/linux/drivers/video/fbdev/omap/ |
H A D | lcd_mipid.c | 153 set_data_lines(md, md->panel.data_lines); in send_init_string() 159 md->hw_guard_end = jiffies + md->hw_guard_wait; in hw_guard_start() 397 md->esd_check(md); in mipid_esd_work() 413 md->enabled = 1; in mipid_enable() 442 md->enabled = 0; in mipid_disable() 470 md->enabled = panel_enabled(md); in mipid_init() 545 md->panel.name, md->revision, md->panel.data_lines); in mipid_detect() 555 md = kzalloc(sizeof(*md), GFP_KERNEL); in mipid_spi_probe() 568 md->spi = spi; in mipid_spi_probe() 581 kfree(md); in mipid_spi_probe() [all …]
|
/linux/drivers/md/ |
H A D | dm.c | 595 io->md = md; in alloc_io() 677 queue_work(md->wq, &md->work); in queue_io() 871 struct mapped_device *md = io->md; in dm_requeue_add_io() local 888 queue_work(md->wq, &md->work); in dm_kick_requeue() 901 struct mapped_device *md = io->md; in dm_handle_requeue() local 945 struct mapped_device *md = io->md; in __dm_io_complete() local 1112 struct mapped_device *md = io->md; in clone_endio() local 1412 struct mapped_device *md = io->md; in __map_bio() local 2140 md->queue = md->disk->queue; in alloc_dev() 2158 md->disk->private_data = md; in alloc_dev() [all …]
|
H A D | dm-era-target.c | 415 md->writeset_tree_info.tm = md->tm; in setup_writeset_tree_info() 451 &md->tm, &md->sm); in create_fresh_metadata() 595 &md->tm, &md->sm); in open_metadata() 821 md->current_writeset = &md->writesets[0]; in metadata_open() 899 r = dm_bitset_flush(&md->bitset_info, md->current_writeset->md.root, in metadata_era_archive() 900 &md->current_writeset->md.root); in metadata_era_archive() 906 ws_pack(&md->current_writeset->md, &value); in metadata_era_archive() 927 &md->writesets[1] : &md->writesets[0]; in next_writeset() 989 r = dm_bitset_flush(&md->bitset_info, md->current_writeset->md.root, in metadata_commit() 990 &md->current_writeset->md.root); in metadata_commit() [all …]
|
H A D | dm-ima.c | 120 dev_name, dev_uuid, md->disk->major, md->disk->first_minor, in dm_ima_alloc_and_copy_device_data() 171 memset(&(md->ima), 0, sizeof(md->ima)); in dm_ima_reset_data() 339 if (table->md->ima.active_table.hash != table->md->ima.inactive_table.hash) in dm_ima_measure_on_table_load() 391 if (md->ima.active_table.hash != md->ima.inactive_table.hash) in dm_ima_measure_on_device_resume() 406 md->ima.active_table.hash = md->ima.inactive_table.hash; in dm_ima_measure_on_device_resume() 407 md->ima.active_table.hash_len = md->ima.inactive_table.hash_len; in dm_ima_measure_on_device_resume() 417 md->ima.active_table.num_targets = md->ima.inactive_table.num_targets; in dm_ima_measure_on_device_resume() 593 if (md->ima.active_table.hash != md->ima.inactive_table.hash) in dm_ima_measure_on_device_remove() 664 md->ima.inactive_table.hash != md->ima.active_table.hash) in dm_ima_measure_on_table_clear() 679 md->ima.inactive_table.hash = md->ima.active_table.hash; in dm_ima_measure_on_table_clear() [all …]
|
H A D | dm-rq.c | 149 dm_put(md); in rq_completed() 160 struct mapped_device *md = tio->md; in dm_end_request() local 190 struct mapped_device *md = tio->md; in dm_requeue_original_request() local 258 struct mapped_device *md = tio->md; in dm_softirq_done() local 341 tio->md = md; in init_tio() 366 struct mapped_device *md = tio->md; in map_request() local 454 dm_get(md); in dm_start_request() 467 tio->md = md; in dm_mq_init_request() 482 struct mapped_device *md = tio->md; in dm_mq_queue_rq() local 547 md->tag_set->numa_node = md->numa_node_id; in dm_mq_init_request_queue() [all …]
|
H A D | dm-ioctl.c | 199 if (!md) in __get_dev_cell() 244 hc->md = md; in alloc_cell() 351 md = hc->md; in dm_hash_remove_all() 504 md = hc->md; in dm_hash_rename() 971 md = hc->md; in find_device() 993 md = hc->md; in dev_remove() 1084 if (!md) in dev_set_geometry() 1127 if (!md) in do_suspend() 1167 md = hc->md; in do_resume() 1250 if (!md) in dev_status() [all …]
|
H A D | dm-zone.c | 25 struct gendisk *disk = md->disk; in dm_blk_do_report_zones() 65 if (dm_suspended_md(md)) in dm_blk_report_zones() 78 if (!md->zone_revalidate_map) in dm_blk_report_zones() 176 md->nr_zones = 0; in dm_revalidate_zones() 178 if (md->nr_zones) in dm_revalidate_zones() 186 md->zone_revalidate_map = t; in dm_revalidate_zones() 226 struct mapped_device *md = t->md; in dm_set_zones_restrictions() local 243 if (!get_capacity(md->disk)) in dm_set_zones_restrictions() 250 md->zone_revalidate_map = t; in dm_set_zones_restrictions() 272 if (!md->disk->nr_zones) { in dm_set_zones_restrictions() [all …]
|
H A D | dm-sysfs.c | 27 struct mapped_device *md; in dm_attr_show() local 34 md = dm_get_from_kobject(kobj); in dm_attr_show() 35 if (!md) in dm_attr_show() 38 ret = dm_attr->show(md, page); in dm_attr_show() 39 dm_put(md); in dm_attr_show() 52 struct mapped_device *md; in dm_attr_store() local 59 md = dm_get_from_kobject(kobj); in dm_attr_store() 60 if (!md) in dm_attr_store() 63 ret = dm_attr->store(md, page, count); in dm_attr_store() 64 dm_put(md); in dm_attr_store() [all …]
|
H A D | dm.h | 77 void dm_lock_md_type(struct mapped_device *md); 78 void dm_unlock_md_type(struct mapped_device *md); 141 int dm_deleting_md(struct mapped_device *md); 146 int dm_suspended_md(struct mapped_device *md); 155 void dm_internal_resume(struct mapped_device *md); 177 int dm_sysfs_init(struct mapped_device *md); 178 void dm_sysfs_exit(struct mapped_device *md); 201 void dm_destroy(struct mapped_device *md); 202 void dm_destroy_immediate(struct mapped_device *md); 203 int dm_open_count(struct mapped_device *md); [all …]
|
/linux/drivers/net/wwan/t7xx/ |
H A D | t7xx_modem_ops.c | 82 struct t7xx_modem *md = t7xx_dev->md; in t7xx_pci_mhccif_isr() local 116 queue_work(md->handshake_wq, &md->handshake_work); in t7xx_pci_mhccif_isr() 589 queue_work(md->handshake_wq, &md->handshake_work); in t7xx_md_event_notify() 596 queue_work(md->handshake_wq, &md->ap_handshake_work); in t7xx_md_event_notify() 649 md = devm_kzalloc(dev, sizeof(*md), GFP_KERNEL); in t7xx_md_alloc() 650 if (!md) in t7xx_md_alloc() 654 t7xx_dev->md = md; in t7xx_md_alloc() 671 return md; in t7xx_md_alloc() 676 struct t7xx_modem *md = t7xx_dev->md; in t7xx_md_reset() local 705 if (!md) in t7xx_md_init() [all …]
|
H A D | t7xx_state_monitor.c | 104 fsm_state_notify(ctl->md, state); in t7xx_fsm_broadcast_state() 228 struct t7xx_modem *md = ctl->md; in t7xx_lk_stage_event_handling() local 234 dev = &md->t7xx_dev->pdev->dev; in t7xx_lk_stage_event_handling() 330 struct t7xx_modem *md = ctl->md; in fsm_routine_ready() local 340 struct t7xx_modem *md = ctl->md; in fsm_routine_starting() local 349 (md->core_md.ready && md->core_ap.ready) || in fsm_routine_starting() 356 if (!md->core_md.ready) { in fsm_routine_starting() 379 struct t7xx_modem *md = ctl->md; in fsm_routine_start() local 384 if (!md) in fsm_routine_start() 623 md->fsm_ctl = ctl; in t7xx_fsm_init() [all …]
|
/linux/drivers/net/mdio/ |
H A D | mdio-mux-bcm-iproc.c | 69 if (md->core_clk) { in mdio_mux_iproc_config() 217 md = devm_kzalloc(&pdev->dev, sizeof(*md), GFP_KERNEL); in mdio_mux_iproc_probe() 218 if (!md) in mdio_mux_iproc_probe() 220 md->dev = &pdev->dev; in mdio_mux_iproc_probe() 223 if (IS_ERR(md->base)) in mdio_mux_iproc_probe() 235 if (!md->mii_bus) { in mdio_mux_iproc_probe() 243 md->core_clk = NULL; in mdio_mux_iproc_probe() 253 bus = md->mii_bus; in mdio_mux_iproc_probe() 254 bus->priv = md; in mdio_mux_iproc_probe() 273 rc = mdio_mux_init(md->dev, md->dev->of_node, mdio_mux_iproc_switch_fn, in mdio_mux_iproc_probe() [all …]
|
H A D | mdio-mux-bcm6368.c | 48 if (md->ext_phy) in bcm6368_mdiomux_read() 69 if (md->ext_phy) in bcm6368_mdiomux_write() 96 md = devm_kzalloc(&pdev->dev, sizeof(*md), GFP_KERNEL); in bcm6368_mdiomux_probe() 97 if (!md) in bcm6368_mdiomux_probe() 110 if (!md->base) { in bcm6368_mdiomux_probe() 116 if (!md->mii_bus) { in bcm6368_mdiomux_probe() 121 bus = md->mii_bus; in bcm6368_mdiomux_probe() 122 bus->priv = md; in bcm6368_mdiomux_probe() 139 rc = mdio_mux_init(md->dev, md->dev->of_node, in bcm6368_mdiomux_probe() 140 bcm6368_mdiomux_switch_fn, &md->mux_handle, md, in bcm6368_mdiomux_probe() [all …]
|
/linux/drivers/soundwire/ |
H A D | master.c | 65 md->bus->prop.clk_freq[i]); in clock_frequencies_show() 107 kfree(md); in sdw_master_device_release() 136 md = kzalloc(sizeof(*md), GFP_KERNEL); in sdw_master_device_add() 137 if (!md) in sdw_master_device_add() 140 md->dev.bus = &sdw_bus_type; in sdw_master_device_add() 142 md->dev.parent = parent; in sdw_master_device_add() 145 md->dev.fwnode = fwnode; in sdw_master_device_add() 157 put_device(&md->dev); in sdw_master_device_add() 162 md->bus = bus; in sdw_master_device_add() 163 bus->dev = &md->dev; in sdw_master_device_add() [all …]
|
/linux/drivers/clk/qcom/ |
H A D | clk-regmap-mux-div.c | 29 val = (div << md->hid_shift) | (src << md->src_shift); in mux_div_set_src_div() 30 mask = ((BIT(md->hid_width) - 1) << md->hid_shift) | in mux_div_set_src_div() 31 ((BIT(md->src_width) - 1) << md->src_shift); in mux_div_set_src_div() 33 ret = regmap_update_bits(md->clkr.regmap, CFG_RCGR + md->reg_offset, in mux_div_set_src_div() 38 ret = regmap_update_bits(md->clkr.regmap, CMD_RCGR + md->reg_offset, in mux_div_set_src_div() 45 ret = regmap_read(md->clkr.regmap, CMD_RCGR + md->reg_offset, in mux_div_set_src_div() 65 regmap_read(md->clkr.regmap, CMD_RCGR + md->reg_offset, &val); in mux_div_get_src_div() 72 regmap_read(md->clkr.regmap, CFG_RCGR + md->reg_offset, &val); in mux_div_get_src_div() 156 md->div = best_div; in __mux_div_set_rate_and_parent() 157 md->src = best_src; in __mux_div_set_rate_and_parent() [all …]
|
/linux/arch/x86/platform/efi/ |
H A D | memmap.c | 181 md = new; in efi_memmap_insert() 183 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_insert() 192 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert() 197 md = new; in efi_memmap_insert() 199 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert() 205 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 210 md = new; in efi_memmap_insert() 218 md = new; in efi_memmap_insert() 227 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 232 md = new; in efi_memmap_insert() [all …]
|
H A D | efi.c | 123 efi_memory_desc_t *md; in do_add_efi_memmap() local 133 switch (md->type) { in do_add_efi_memmap() 183 efi_memory_desc_t *md; in do_efi_soft_reserve() local 247 u64 end = (md->num_pages << EFI_PAGE_SHIFT) + md->phys_addr - 1; in efi_memmap_entry_valid() 365 md->phys_addr, in efi_print_memmap() 366 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1, in efi_print_memmap() 541 prev_md = md; in efi_merge_regions() 547 prev_md = md; in efi_merge_regions() 556 md->attribute = 0; in efi_merge_regions() 559 prev_md = md; in efi_merge_regions() [all …]
|
/linux/arch/arm64/kernel/ |
H A D | efi.c | 22 return !PAGE_ALIGNED(md->phys_addr) || in region_is_misaligned() 33 u64 attr = md->attribute; in create_mapping_protection() 34 u32 type = md->type; in create_mapping_protection() 39 if (region_is_misaligned(md)) { in create_mapping_protection() 89 if (region_is_misaligned(md)) in efi_create_mapping() 92 create_pgd_mapping(mm, md->phys_addr, md->virt_addr, in efi_create_mapping() 99 const efi_memory_desc_t *md; member 106 const efi_memory_desc_t *md = spd->md; in set_permissions() local 109 if (md->attribute & EFI_MEMORY_RO) in set_permissions() 120 efi_memory_desc_t *md, in efi_set_mapping_permissions() argument [all …]
|
/linux/drivers/rapidio/devices/ |
H A D | rio_mport_cdev.c | 385 map->md = md; in rio_mport_create_outbound_mapping() 464 struct mport_dev *md = priv->md; in rio_mport_obw_free() local 499 struct mport_dev *md = priv->md; in maint_hdid_set() local 521 struct mport_dev *md = priv->md; in maint_comptag_set() local 816 struct mport_dev *md = priv->md; in rio_dma_transfer() local 1107 map->md = md; in rio_mport_create_dma_mapping() 1120 struct mport_dev *md = priv->md; in rio_mport_alloc_dma() local 1147 struct mport_dev *md = priv->md; in rio_mport_free_dma() local 1234 map->md = md; in rio_mport_create_inbound_mapping() 1965 md = priv->md; in mport_cdev_release_dma() [all …]
|
/linux/drivers/dma/ |
H A D | uniphier-mdmac.c | 114 sg = &md->sgl[md->sg_cur]; in uniphier_mdmac_handle() 150 if (md) in uniphier_mdmac_start() 203 md = mc->md; in uniphier_mdmac_interrupt() 204 if (!md) in uniphier_mdmac_interrupt() 209 if (md->sg_cur >= md->sg_len) { in uniphier_mdmac_interrupt() 212 if (!md) in uniphier_mdmac_interrupt() 241 md = kzalloc(sizeof(*md), GFP_NOWAIT); in uniphier_mdmac_prep_slave_sg() 242 if (!md) in uniphier_mdmac_prep_slave_sg() 304 if (mc->md && mc->md->vd.tx.cookie == cookie) { in uniphier_mdmac_tx_status() 308 md = mc->md; in uniphier_mdmac_tx_status() [all …]
|
H A D | milbeaut-hdmac.c | 122 sg = &md->sgl[md->sg_cur]; in milbeaut_chan_start() 167 if (md) in milbeaut_hdmac_start() 187 md = mc->md; in milbeaut_hdmac_interrupt() 188 if (!md) in milbeaut_hdmac_interrupt() 193 if (md->sg_cur >= md->sg_len) { in milbeaut_hdmac_interrupt() 268 md = kzalloc(sizeof(*md), GFP_NOWAIT); in milbeaut_hdmac_prep_slave_sg() 269 if (!md) in milbeaut_hdmac_prep_slave_sg() 345 if (mc->md && mc->md->vd.tx.cookie == cookie) { in milbeaut_hdmac_tx_status() 349 md = mc->md; in milbeaut_hdmac_tx_status() 350 sg = &md->sgl[md->sg_cur]; in milbeaut_hdmac_tx_status() [all …]
|
/linux/arch/riscv/kernel/ |
H A D | efi.c | 21 u64 attr = md->attribute; in efimem_to_pgprot_map() 22 u32 type = md->type; in efimem_to_pgprot_map() 53 for (i = 0; i < md->num_pages; i++) in efi_create_mapping() 55 md->phys_addr + i * PAGE_SIZE, in efi_create_mapping() 62 efi_memory_desc_t *md = data; in set_permissions() local 66 if (md->attribute & EFI_MEMORY_RO) { in set_permissions() 71 if (md->attribute & EFI_MEMORY_XP) { in set_permissions() 81 efi_memory_desc_t *md, in efi_set_mapping_permissions() argument 94 return apply_to_page_range(mm, md->virt_addr, in efi_set_mapping_permissions() 95 md->num_pages << EFI_PAGE_SHIFT, in efi_set_mapping_permissions() [all …]
|
/linux/drivers/firmware/efi/ |
H A D | efi-init.c | 41 efi_memory_desc_t *md; in efi_to_phys() local 43 for_each_efi_memory_desc(md) { in efi_to_phys() 46 if (md->virt_addr == 0) in efi_to_phys() 49 if (md->virt_addr <= addr && in efi_to_phys() 50 (addr - md->virt_addr) < (md->num_pages << EFI_PAGE_SHIFT)) in efi_to_phys() 51 return md->phys_addr + addr - md->virt_addr; in efi_to_phys() 138 switch (md->type) { in is_usable_memory() 160 efi_memory_desc_t *md; in reserve_regions() local 175 paddr = md->phys_addr; in reserve_regions() 176 npages = md->num_pages; in reserve_regions() [all …]
|
/linux/tools/thermal/lib/ |
H A D | mainloop.c | 26 struct mainloop_data *md; in mainloop() local 45 md = events[i].data.ptr; in mainloop() 47 if (md->cb(md->fd, md->data) > 0) in mainloop() 59 struct mainloop_data *md; in mainloop_add() local 61 md = malloc(sizeof(*md)); in mainloop_add() 62 if (!md) in mainloop_add() 65 md->data = data; in mainloop_add() 66 md->cb = cb; in mainloop_add() 67 md->fd = fd; in mainloop_add() 69 ev.data.ptr = md; in mainloop_add() [all …]
|
/linux/drivers/firmware/efi/libstub/ |
H A D | randomalloc.c | 25 if (md->type != EFI_CONVENTIONAL_MEMORY) in get_entry_num_slots() 29 (md->attribute & EFI_MEMORY_SP)) in get_entry_num_slots() 32 region_end = min(md->phys_addr + md->num_pages * EFI_PAGE_SIZE - 1, in get_entry_num_slots() 52 #define MD_NUM_SLOTS(md) ((md)->virt_addr) argument 79 efi_memory_desc_t *md = (void *)map->map + map_offset; in efi_random_alloc() local 84 MD_NUM_SLOTS(md) = slots; in efi_random_alloc() 86 if (md->attribute & EFI_MEMORY_MORE_RELIABLE) in efi_random_alloc() 110 efi_memory_desc_t *md = (void *)map->map + map_offset; in efi_random_alloc() local 115 !(md->attribute & EFI_MEMORY_MORE_RELIABLE)) in efi_random_alloc() 118 if (target_slot >= MD_NUM_SLOTS(md)) { in efi_random_alloc() [all …]
|