/linux-master/drivers/md/ |
H A D | dm-ima.h | 59 void dm_ima_reset_data(struct mapped_device *md); 61 void dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap); 62 void dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all); 63 void dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map); 64 void dm_ima_measure_on_device_rename(struct mapped_device *md); 68 static inline void dm_ima_reset_data(struct mapped_device *md) {} argument 70 static inline void dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap) {} argument 71 static inline void dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all) {} argument 72 static inline void dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map) {} argument 73 static inline void dm_ima_measure_on_device_rename(struct mapped_device *md) {} argument [all...] |
H A D | dm.c | 308 int dm_deleting_md(struct mapped_device *md) argument 310 return test_bit(DMF_DELETING, &md->flags); 315 struct mapped_device *md; local 319 md = disk->private_data; 320 if (!md) 323 if (test_bit(DMF_FREEING, &md->flags) || 324 dm_deleting_md(md)) { 325 md = NULL; 329 dm_get(md); 330 atomic_inc(&md 339 struct mapped_device *md; local 356 dm_open_count(struct mapped_device *md) argument 364 dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred) argument 384 dm_cancel_deferred_remove(struct mapped_device *md) argument 407 struct mapped_device *md = bdev->bd_disk->private_data; local 412 dm_prepare_ioctl(struct mapped_device *md, int *srcu_idx, struct block_device **bdev) argument 446 dm_unprepare_ioctl(struct mapped_device *md, int srcu_idx) argument 454 struct mapped_device *md = bdev->bd_disk->private_data; local 573 alloc_io(struct mapped_device *md, struct bio *bio, gfp_t gfp_mask) argument 617 struct mapped_device *md = ci->io->md; local 670 queue_io(struct mapped_device *md, struct bio *bio) argument 699 dm_sync_table(struct mapped_device *md) argument 725 open_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode) argument 777 close_table_device(struct table_device *td, struct mapped_device *md) argument 805 dm_get_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode, struct dm_dev **result) argument 827 dm_put_table_device(struct mapped_device *md, struct dm_dev *d) argument 840 dm_get_geometry(struct mapped_device *md, struct hd_geometry *geo) argument 850 dm_set_geometry(struct mapped_device *md, struct hd_geometry *geo) argument 864 __noflush_suspending(struct mapped_device *md) argument 871 struct mapped_device *md = io->md; local 883 dm_kick_requeue(struct mapped_device *md, bool first_stage) argument 901 struct mapped_device *md = io->md; local 945 struct mapped_device *md = io->md; local 993 struct mapped_device *md = container_of(work, struct mapped_device, local 1079 dm_get_queue_limits(struct mapped_device *md) argument 1084 disable_discard(struct mapped_device *md) argument 1092 disable_write_zeroes(struct mapped_device *md) argument 1112 struct mapped_device *md = io->md; local 1236 struct mapped_device *md = dax_get_private(dax_dev); local 1263 struct mapped_device *md = dax_get_private(dax_dev); local 1290 struct mapped_device *md = dax_get_private(dax_dev); local 1391 __set_swap_bios_limit(struct mapped_device *md, int latch) argument 1412 struct mapped_device *md = io->md; local 1765 dm_zone_bio_needs_split(struct mapped_device *md, struct bio *bio) argument 1775 dm_zone_plug_bio(struct mapped_device *md, struct bio *bio) argument 1780 dm_zone_bio_needs_split(struct mapped_device *md, struct bio *bio) argument 1785 dm_zone_plug_bio(struct mapped_device *md, struct bio *bio) argument 1794 dm_split_and_process_bio(struct mapped_device *md, struct dm_table *map, struct bio *bio) argument 1878 struct mapped_device *md = bio->bi_bdev->bd_disk->private_data; local 2033 cleanup_mapped_device(struct mapped_device *md) argument 2093 struct mapped_device *md; local 2209 free_dev(struct mapped_device *md) argument 2232 struct mapped_device *md = context; local 2248 __bind(struct mapped_device *md, struct dm_table *t, struct queue_limits *limits) argument 2318 __unbind(struct mapped_device *md) argument 2337 struct mapped_device *md; local 2353 dm_lock_md_type(struct mapped_device *md) argument 2358 dm_unlock_md_type(struct mapped_device *md) argument 2363 dm_set_md_type(struct mapped_device *md, enum dm_queue_mode type) argument 2369 dm_get_md_type(struct mapped_device *md) argument 2374 dm_get_immutable_target_type(struct mapped_device *md) argument 2382 dm_setup_md_queue(struct mapped_device *md, struct dm_table *t) argument 2454 struct mapped_device *md; local 2476 dm_get_mdptr(struct mapped_device *md) argument 2481 dm_set_mdptr(struct mapped_device *md, void *ptr) argument 2486 dm_get(struct mapped_device *md) argument 2492 dm_hold(struct mapped_device *md) argument 2505 dm_device_name(struct mapped_device *md) argument 2511 __dm_destroy(struct mapped_device *md, bool wait) argument 2558 dm_destroy(struct mapped_device *md) argument 2563 dm_destroy_immediate(struct mapped_device *md) argument 2568 dm_put(struct mapped_device *md) argument 2574 dm_in_flight_bios(struct mapped_device *md) argument 2585 dm_wait_for_bios_completion(struct mapped_device *md, unsigned int task_state) argument 2610 dm_wait_for_completion(struct mapped_device *md, unsigned int task_state) argument 2637 struct mapped_device *md = container_of(work, struct mapped_device, work); local 2653 dm_queue_flush(struct mapped_device *md) argument 2663 dm_swap_table(struct mapped_device *md, struct dm_table *table) argument 2708 lock_fs(struct mapped_device *md) argument 2720 unlock_fs(struct mapped_device *md) argument 2737 __dm_suspend(struct mapped_device *md, struct dm_table *map, unsigned int suspend_flags, unsigned int task_state, int dmf_suspended_flag) argument 2845 dm_suspend(struct mapped_device *md, unsigned int suspend_flags) argument 2886 __dm_resume(struct mapped_device *md, struct dm_table *map) argument 2910 dm_resume(struct mapped_device *md) argument 2952 __dm_internal_suspend(struct mapped_device *md, unsigned int suspend_flags) argument 2982 __dm_internal_resume(struct mapped_device *md) argument 3018 dm_internal_suspend_noflush(struct mapped_device *md) argument 3026 dm_internal_resume(struct mapped_device *md) argument 3039 dm_internal_suspend_fast(struct mapped_device *md) argument 3052 dm_internal_resume_fast(struct mapped_device *md) argument 3069 dm_kobject_uevent(struct mapped_device *md, enum kobject_action action, unsigned int cookie, bool need_resize_uevent) argument 3095 dm_next_uevent_seq(struct mapped_device *md) argument 3100 dm_get_event_nr(struct mapped_device *md) argument 3105 dm_wait_event(struct mapped_device *md, int event_nr) argument 3111 dm_uevent_add(struct mapped_device *md, struct list_head *elist) argument 3124 dm_disk(struct mapped_device *md) argument 3130 dm_kobject(struct mapped_device *md) argument 3137 struct mapped_device *md; local 3153 dm_suspended_md(struct mapped_device *md) argument 3158 dm_post_suspending_md(struct mapped_device *md) argument 3163 dm_suspended_internally_md(struct mapped_device *md) argument 3168 dm_test_deferred_remove_flag(struct mapped_device *md) argument 3217 struct mapped_device *md = bdev->bd_disk->private_data; local 3426 struct mapped_device *md = bdev->bd_disk->private_data; local [all...] |
H A D | dm-sysfs.c | 15 ssize_t (*show)(struct mapped_device *md, char *p); 16 ssize_t (*store)(struct mapped_device *md, const char *p, size_t count); 27 struct mapped_device *md; local 34 md = dm_get_from_kobject(kobj); 35 if (!md) 38 ret = dm_attr->show(md, page); 39 dm_put(md); 52 struct mapped_device *md; local 59 md = dm_get_from_kobject(kobj); 60 if (!md) 69 dm_attr_name_show(struct mapped_device *md, char *buf) argument 78 dm_attr_uuid_show(struct mapped_device *md, char *buf) argument 87 dm_attr_suspended_show(struct mapped_device *md, char *buf) argument 94 dm_attr_use_blk_mq_show(struct mapped_device *md, char *buf) argument 133 dm_sysfs_init(struct mapped_device *md) argument 143 dm_sysfs_exit(struct mapped_device *md) argument [all...] |
H A D | dm-era-target.c | 36 struct writeset_metadata md; member in struct:writeset 96 ws->md.nr_bits = nr_blocks; 97 r = setup_on_disk_bitset(info, ws->md.nr_bits, &ws->md.root); 142 r = dm_bitset_set_bit(info, ws->md.root, block, &ws->md.root); 311 static int superblock_read_lock(struct era_metadata *md, argument 314 return dm_bm_read_lock(md->bm, SUPERBLOCK_LOCATION, 318 static int superblock_lock_zero(struct era_metadata *md, argument 321 return dm_bm_write_lock_zero(md 325 superblock_lock(struct era_metadata *md, struct dm_block **sblock) argument 378 struct era_metadata *md = context; local 392 struct era_metadata *md = context; local 411 setup_writeset_tree_info(struct era_metadata *md) argument 424 setup_era_array_info(struct era_metadata *md) argument 437 setup_infos(struct era_metadata *md) argument 446 create_fresh_metadata(struct era_metadata *md) argument 480 save_sm_root(struct era_metadata *md) argument 493 copy_sm_root(struct era_metadata *md, struct superblock_disk *disk) argument 505 prepare_superblock(struct era_metadata *md, struct superblock_disk *disk) argument 527 write_superblock(struct era_metadata *md) argument 552 format_metadata(struct era_metadata *md) argument 570 open_metadata(struct era_metadata *md) argument 621 open_or_format_metadata(struct era_metadata *md, bool may_format) argument 637 create_persistent_data_objects(struct era_metadata *md, bool may_format) argument 656 destroy_persistent_data_objects(struct era_metadata *md) argument 666 swap_writeset(struct era_metadata *md, struct writeset *new_writeset) argument 694 metadata_digest_remove_writeset(struct era_metadata *md, struct digest *d) argument 713 metadata_digest_transcribe_writeset(struct era_metadata *md, struct digest *d) argument 747 metadata_digest_lookup_writeset(struct era_metadata *md, struct digest *d) argument 789 metadata_digest_start(struct era_metadata *md, struct digest *d) argument 811 struct era_metadata *md = kzalloc(sizeof(*md), GFP_KERNEL); local 832 metadata_close(struct era_metadata *md) argument 849 metadata_resize(struct era_metadata *md, void *arg) argument 893 metadata_era_archive(struct era_metadata *md) argument 924 next_writeset(struct era_metadata *md) argument 930 metadata_new_era(struct era_metadata *md) argument 947 metadata_era_rollover(struct era_metadata *md) argument 970 metadata_current_marked(struct era_metadata *md, dm_block_t block) argument 983 metadata_commit(struct era_metadata *md) argument 1020 metadata_checkpoint(struct era_metadata *md) argument 1032 metadata_take_snap(struct era_metadata *md) argument 1091 metadata_drop_snap(struct era_metadata *md) argument 1144 metadata_get_stats(struct era_metadata *md, void *ptr) argument 1181 struct era_metadata *md; member in struct:era 1393 in_worker0(struct era *era, int (*fn)(struct era_metadata *md)) argument 1403 in_worker1(struct era *era, int (*fn)(struct era_metadata *md, void *ref), void *arg) argument 1469 struct era_metadata *md; local [all...] |
H A D | dm-rq.h | 34 int dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t); 35 void dm_mq_cleanup_mapped_device(struct mapped_device *md); 40 void dm_mq_kick_requeue_list(struct mapped_device *md); 44 ssize_t dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf); 45 ssize_t dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md,
|
H A D | dm-rq.c | 19 struct mapped_device *md; member in struct:dm_rq_target_io 59 int dm_request_based(struct mapped_device *md) argument 61 return queue_is_mq(md->queue); 127 static void rq_end_stats(struct mapped_device *md, struct request *orig) argument 129 if (unlikely(dm_stats_used(&md->stats))) { 133 dm_stats_account_io(&md->stats, rq_data_dir(orig), 140 * Don't touch any member of the md after calling this function because 141 * the md may be freed in dm_put() at the end of this function. 144 static void rq_completed(struct mapped_device *md) argument 149 dm_put(md); 160 struct mapped_device *md = tio->md; local 176 dm_mq_kick_requeue_list(struct mapped_device *md) argument 190 struct mapped_device *md = tio->md; local 258 struct mapped_device *md = tio->md; local 338 init_tio(struct dm_rq_target_io *tio, struct request *rq, struct mapped_device *md) argument 366 struct mapped_device *md = tio->md; local 422 dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf) argument 427 dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md, const char *buf, size_t count) argument 433 dm_start_request(struct mapped_device *md, struct request *orig) argument 460 struct mapped_device *md = set->driver_data; local 482 struct mapped_device *md = tio->md; local 536 dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t) argument 578 dm_mq_cleanup_mapped_device(struct mapped_device *md) argument [all...] |
H A D | dm-ima.c | 67 static int dm_ima_alloc_and_copy_name_uuid(struct mapped_device *md, char **dev_name, argument 83 r = dm_copy_name_and_uuid(md, *dev_name, *dev_uuid); 102 static int dm_ima_alloc_and_copy_device_data(struct mapped_device *md, char **device_data, argument 108 r = dm_ima_alloc_and_copy_name_uuid(md, &dev_name, &dev_uuid, noio); 120 dev_name, dev_uuid, md->disk->major, md->disk->first_minor, 121 md->disk->minors, num_targets); 149 static int dm_ima_alloc_and_copy_capacity_str(struct mapped_device *md, char **capacity_str, argument 154 capacity = get_capacity(md->disk); 169 void dm_ima_reset_data(struct mapped_device *md) argument 369 dm_ima_measure_on_device_resume(struct mapped_device *md, bool swap) argument 474 dm_ima_measure_on_device_remove(struct mapped_device *md, bool remove_all) argument 605 dm_ima_measure_on_table_clear(struct mapped_device *md, bool new_map) argument 704 dm_ima_measure_on_device_rename(struct mapped_device *md) argument [all...] |
H A D | dm-zone.c | 21 static int dm_blk_do_report_zones(struct mapped_device *md, struct dm_table *t, argument 25 struct gendisk *disk = md->disk; 59 struct mapped_device *md = disk->private_data; local 63 if (!md->zone_revalidate_map) { 65 if (dm_suspended_md(md)) 68 map = dm_get_live_table(md, &srcu_idx); 73 map = md->zone_revalidate_map; 76 ret = dm_blk_do_report_zones(md, map, sector, nr_zones, cb, data); 78 if (!md->zone_revalidate_map) 79 dm_put_live_table(md, srcu_id 132 dm_is_zone_write(struct mapped_device *md, struct bio *bio) argument 169 dm_revalidate_zones(struct mapped_device *md, struct dm_table *t) argument 226 struct mapped_device *md = t->md; local 292 struct mapped_device *md = io->md; local [all...] |
H A D | dm.h | 77 void dm_lock_md_type(struct mapped_device *md); 78 void dm_unlock_md_type(struct mapped_device *md); 79 void dm_set_md_type(struct mapped_device *md, enum dm_queue_mode type); 80 enum dm_queue_mode dm_get_md_type(struct mapped_device *md); 81 struct target_type *dm_get_immutable_target_type(struct mapped_device *md); 83 int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t); 110 bool dm_is_zone_write(struct mapped_device *md, struct bio *bio); 114 static inline bool dm_is_zone_write(struct mapped_device *md, struct bio *bio) argument 141 int dm_deleting_md(struct mapped_device *md); 146 int dm_suspended_md(struct mapped_device *md); [all...] |
/linux-master/drivers/video/fbdev/omap/ |
H A D | lcd_mipid.c | 56 static void mipid_transfer(struct mipid_device *md, int cmd, const u8 *wbuf, argument 64 BUG_ON(md->spi == NULL); 105 r = spi_sync(md->spi, &m); 107 dev_dbg(&md->spi->dev, "spi_sync %d\n", r); 113 static inline void mipid_cmd(struct mipid_device *md, int cmd) argument 115 mipid_transfer(md, cmd, NULL, 0, NULL, 0); 118 static inline void mipid_write(struct mipid_device *md, argument 121 mipid_transfer(md, reg, buf, len, NULL, 0); 124 static inline void mipid_read(struct mipid_device *md, argument 127 mipid_transfer(md, re 130 set_data_lines(struct mipid_device *md, int data_lines) argument 148 send_init_string(struct mipid_device *md) argument 156 hw_guard_start(struct mipid_device *md, int guard_msec) argument 162 hw_guard_wait(struct mipid_device *md) argument 172 set_sleep_mode(struct mipid_device *md, int on) argument 194 set_display_state(struct mipid_device *md, int enabled) argument 203 struct mipid_device *md = to_mipid_device(panel); local 221 struct mipid_device *md = to_mipid_device(panel); local 231 struct mipid_device *md = to_mipid_device(panel); local 245 read_first_pixel(struct mipid_device *md) argument 275 struct mipid_device *md = to_mipid_device(panel); local 312 ls041y3_esd_recover(struct mipid_device *md) argument 319 ls041y3_esd_check_mode1(struct mipid_device *md) argument 335 ls041y3_esd_check_mode2(struct mipid_device *md) argument 371 ls041y3_esd_check(struct mipid_device *md) argument 378 mipid_esd_start_check(struct mipid_device *md) argument 385 mipid_esd_stop_check(struct mipid_device *md) argument 393 struct mipid_device *md = container_of(work, struct mipid_device, local 404 struct mipid_device *md = to_mipid_device(panel); local 425 struct mipid_device *md = to_mipid_device(panel); local 447 panel_enabled(struct mipid_device *md) argument 464 struct mipid_device *md = to_mipid_device(panel); local 482 struct mipid_device *md = to_mipid_device(panel); local 513 mipid_detect(struct mipid_device *md) argument 552 struct mipid_device *md; local 587 struct mipid_device *md = dev_get_drvdata(&spi->dev); local [all...] |
/linux-master/tools/thermal/lib/ |
H A D | mainloop.c | 26 struct mainloop_data *md; local 45 md = events[i].data.ptr; 47 if (md->cb(md->fd, md->data) > 0) 59 struct mainloop_data *md; local 61 md = malloc(sizeof(*md)); 62 if (!md) 65 md [all...] |
/linux-master/drivers/soundwire/ |
H A D | master.c | 42 struct sdw_master_device *md = dev_to_sdw_master_device(dev); \ 43 return sprintf(buf, format_string, md->bus->prop.field); \ 59 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local 63 for (i = 0; i < md->bus->prop.num_clk_freq; i++) 65 md->bus->prop.clk_freq[i]); 75 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local 79 for (i = 0; i < md->bus->prop.num_clk_gears; i++) 81 md->bus->prop.clk_gears[i]); 105 struct sdw_master_device *md = dev_to_sdw_master_device(dev); local 107 kfree(md); 130 struct sdw_master_device *md; local [all...] |
/linux-master/arch/riscv/kernel/ |
H A D | efi.c | 19 static __init pgprot_t efimem_to_pgprot_map(efi_memory_desc_t *md) argument 21 u64 attr = md->attribute; 22 u32 type = md->type; 46 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument 48 pgprot_t prot = __pgprot(pgprot_val(efimem_to_pgprot_map(md)) & 53 for (i = 0; i < md->num_pages; i++) 54 create_pgd_mapping(mm->pgd, md->virt_addr + i * PAGE_SIZE, 55 md->phys_addr + i * PAGE_SIZE, 62 efi_memory_desc_t *md = data; local 66 if (md 80 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool ignored) argument [all...] |
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_modem_ops.c | 82 struct t7xx_modem *md = t7xx_dev->md; local 88 ctl = md->fsm_ctl; 95 spin_lock_bh(&md->exp_lock); 97 md->exp_id |= int_sta; 98 if (md->exp_id & D2H_INT_EXCEPTION_INIT) { 103 md->exp_id &= ~D2H_INT_EXCEPTION_INIT; 106 } else if (md->exp_id & D2H_INT_PORT_ENUM) { 107 md->exp_id &= ~D2H_INT_PORT_ENUM; 114 if ((md 297 t7xx_md_exception(struct t7xx_modem *md, enum hif_ex_stage stage) argument 316 t7xx_wait_hif_ex_hk_event(struct t7xx_modem *md, int event_id) argument 444 t7xx_core_reset(struct t7xx_modem *md) argument 467 t7xx_core_hk_handler(struct t7xx_modem *md, struct t7xx_sys_info *core_info, struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_event_state event_id, enum t7xx_fsm_event_state err_detect) argument 533 struct t7xx_modem *md = container_of(work, struct t7xx_modem, handshake_work); local 547 struct t7xx_modem *md = container_of(work, struct t7xx_modem, ap_handshake_work); local 559 t7xx_md_event_notify(struct t7xx_modem *md, enum md_event_id evt_id) argument 621 t7xx_md_exception_handshake(struct t7xx_modem *md) argument 647 struct t7xx_modem *md; local 676 struct t7xx_modem *md = t7xx_dev->md; local 701 struct t7xx_modem *md; local 768 struct t7xx_modem *md = t7xx_dev->md; local [all...] |
/linux-master/drivers/net/mdio/ |
H A D | mdio-mux-bcm6368.c | 39 struct bcm6368_mdiomux_desc *md = bus->priv; local 43 __raw_writel(0, md->base + MDIOC_REG); 48 if (md->ext_phy) 51 __raw_writel(reg, md->base + MDIOC_REG); 53 ret = __raw_readw(md->base + MDIOD_REG); 61 struct bcm6368_mdiomux_desc *md = bus->priv; local 64 __raw_writel(0, md->base + MDIOC_REG); 69 if (md->ext_phy) 73 __raw_writel(reg, md->base + MDIOC_REG); 82 struct bcm6368_mdiomux_desc *md local 91 struct bcm6368_mdiomux_desc *md; local 158 struct bcm6368_mdiomux_desc *md = platform_get_drvdata(pdev); local [all...] |
H A D | mdio-mux-bcm-iproc.c | 59 static void mdio_mux_iproc_config(struct iproc_mdiomux_desc *md) argument 65 val = readl(md->base + MDIO_SCAN_CTRL_OFFSET); 67 writel(val, md->base + MDIO_SCAN_CTRL_OFFSET); 69 if (md->core_clk) { 73 divisor = clk_get_rate(md->core_clk) / MDIO_OPERATING_FREQUENCY; 77 writel(val, md->base + MDIO_RATE_ADJ_EXT_OFFSET); 78 writel(val, md->base + MDIO_RATE_ADJ_INT_OFFSET); 138 struct iproc_mdiomux_desc *md = bus->priv; local 141 ret = start_miim_ops(md->base, false, phyid, reg, 0, MDIO_CTRL_READ_OP); 151 struct iproc_mdiomux_desc *md local 165 struct iproc_mdiomux_desc *md = bus->priv; local 180 struct iproc_mdiomux_desc *md = bus->priv; local 195 struct iproc_mdiomux_desc *md = data; local 212 struct iproc_mdiomux_desc *md; local 294 struct iproc_mdiomux_desc *md = platform_get_drvdata(pdev); local 304 struct iproc_mdiomux_desc *md = dev_get_drvdata(dev); local 313 struct iproc_mdiomux_desc *md = dev_get_drvdata(dev); local [all...] |
/linux-master/drivers/clk/qcom/ |
H A D | clk-regmap-mux-div.c | 23 int mux_div_set_src_div(struct clk_regmap_mux_div *md, u32 src, u32 div) argument 27 const char *name = clk_hw_get_name(&md->clkr.hw); 29 val = (div << md->hid_shift) | (src << md->src_shift); 30 mask = ((BIT(md->hid_width) - 1) << md->hid_shift) | 31 ((BIT(md->src_width) - 1) << md->src_shift); 33 ret = regmap_update_bits(md->clkr.regmap, CFG_RCGR + md 59 mux_div_get_src_div(struct clk_regmap_mux_div *md, u32 *src, u32 *div) argument 91 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 127 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 165 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 181 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 189 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 197 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local 205 struct clk_regmap_mux_div *md = to_clk_regmap_mux_div(hw); local [all...] |
/linux-master/arch/arm64/kernel/ |
H A D | efi.c | 18 static bool region_is_misaligned(const efi_memory_desc_t *md) argument 22 return !PAGE_ALIGNED(md->phys_addr) || 23 !PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT); 31 static __init pteval_t create_mapping_protection(efi_memory_desc_t *md) argument 33 u64 attr = md->attribute; 34 u32 type = md->type; 39 if (region_is_misaligned(md)) { 75 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument 77 pteval_t prot_val = create_mapping_protection(md); 78 bool page_mappings_only = (md 99 const efi_memory_desc_t *md; member in struct:set_perm_data 106 const efi_memory_desc_t *md = spd->md; local 119 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool has_bti) argument [all...] |
/linux-master/drivers/firmware/efi/libstub/ |
H A D | randomalloc.c | 17 static unsigned long get_entry_num_slots(efi_memory_desc_t *md, argument 25 if (md->type != EFI_CONVENTIONAL_MEMORY) 29 (md->attribute & EFI_MEMORY_SP)) 32 region_end = min(md->phys_addr + md->num_pages * EFI_PAGE_SIZE - 1, 37 first_slot = round_up(max(md->phys_addr, alloc_min), align); 52 #define MD_NUM_SLOTS(md) ((md)->virt_addr) 79 efi_memory_desc_t *md = (void *)map->map + map_offset; local 82 slots = get_entry_num_slots(md, siz 110 efi_memory_desc_t *md = (void *)map->map + map_offset; local [all...] |
/linux-master/tools/build/feature/ |
H A D | test-libcrypto.c | 9 unsigned char md[MD5_DIGEST_LENGTH + SHA_DIGEST_LENGTH]; local 19 EVP_DigestFinal_ex(mdctx, &md[0], &digest_len); 22 SHA1(&dat[0], sizeof(dat), &md[0]);
|
/linux-master/drivers/firmware/efi/ |
H A D | memattr.c | 60 efi_memory_desc_t *md; local 84 for_each_efi_memory_desc(md) { 85 u64 md_paddr = md->phys_addr; 86 u64 md_size = md->num_pages << EFI_PAGE_SHIFT; 88 if (!(md->attribute & EFI_MEMORY_RUNTIME)) 90 if (md->virt_addr == 0 && md->phys_addr != 0) { 107 if (md->type != in->type) { 112 out->virt_addr = in_paddr + (md->virt_addr - md_paddr); 162 efi_memory_desc_t md; local [all...] |
H A D | efi-init.c | 27 static int __init is_memory(efi_memory_desc_t *md) argument 29 if (md->attribute & (EFI_MEMORY_WB|EFI_MEMORY_WT|EFI_MEMORY_WC)) 41 efi_memory_desc_t *md; local 43 for_each_efi_memory_desc(md) { 44 if (!(md->attribute & EFI_MEMORY_RUNTIME)) 46 if (md->virt_addr == 0) 49 if (md->virt_addr <= addr && 50 (addr - md->virt_addr) < (md->num_pages << EFI_PAGE_SHIFT)) 51 return md 136 is_usable_memory(efi_memory_desc_t *md) argument 160 efi_memory_desc_t *md; local [all...] |
/linux-master/arch/x86/platform/efi/ |
H A D | efi_32.c | 36 void __init efi_map_region(efi_memory_desc_t *md) argument 42 start_pfn = PFN_DOWN(md->phys_addr); 43 size = md->num_pages << PAGE_SHIFT; 44 end = md->phys_addr + size; 48 va = __va(md->phys_addr); 50 if (!(md->attribute & EFI_MEMORY_WB)) 51 set_memory_uc((unsigned long)va, md->num_pages); 53 va = ioremap_cache(md->phys_addr, size); 56 md->virt_addr = (unsigned long)va; 58 pr_err("ioremap of 0x%llX failed!\n", md 87 efi_map_region_fixed(efi_memory_desc_t *md) argument 132 efi_memory_desc_t *md; local [all...] |
H A D | memmap.c | 105 * @md: EFI memory descriptor to split 111 int __init efi_memmap_split_count(efi_memory_desc_t *md, struct range *range) argument 117 start = md->phys_addr; 118 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; 155 efi_memory_desc_t *md; local 181 md = new; 182 start = md->phys_addr; 183 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; 186 md [all...] |
/linux-master/arch/arm/kernel/ |
H A D | efi.c | 16 efi_memory_desc_t *md = data; local 19 if (md->attribute & EFI_MEMORY_RO) 21 if (md->attribute & EFI_MEMORY_XP) 28 efi_memory_desc_t *md, 33 base = md->virt_addr; 34 size = md->num_pages << EFI_PAGE_SHIFT; 44 return apply_to_page_range(mm, base, size, set_permissions, md); 49 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) argument 52 .virtual = md->virt_addr, 53 .pfn = __phys_to_pfn(md 27 efi_set_mapping_permissions(struct mm_struct *mm, efi_memory_desc_t *md, bool ignored) argument [all...] |