if (!res || resource_type(res) != IORESOURCE_MEM) {
ret = dev_err_probe(dev, -EINVAL, "invalid resource %pR\n", res); return IOMEM_ERR_PTR(ret);
}
if (type == DEVM_IOREMAP && res->flags & IORESOURCE_MEM_NONPOSTED)
type = DEVM_IOREMAP_NP;
size = resource_size(res);
if (res->name)
pretty_name = devm_kasprintf(dev, GFP_KERNEL, "%s %s",
dev_name(dev), res->name); else
pretty_name = devm_kstrdup(dev, dev_name(dev), GFP_KERNEL); if (!pretty_name) {
ret = dev_err_probe(dev, -ENOMEM, "can't generate pretty name for resource %pR\n", res); return IOMEM_ERR_PTR(ret);
}
if (!devm_request_mem_region(dev, res->start, size, pretty_name)) {
ret = dev_err_probe(dev, -EBUSY, "can't request region for resource %pR\n", res); return IOMEM_ERR_PTR(ret);
}
dest_ptr = __devm_ioremap(dev, res->start, size, type); if (!dest_ptr) {
devm_release_mem_region(dev, res->start, size);
ret = dev_err_probe(dev, -ENOMEM, "ioremap failed for resource %pR\n", res); return IOMEM_ERR_PTR(ret);
}
return dest_ptr;
}
/** * devm_ioremap_resource() - check, request region, and ioremap resource * @dev: generic device to handle the resource for * @res: resource to be handled * * Checks that a resource is a valid memory region, requests the memory * region and ioremaps it. All operations are managed and will be undone * on driver detach. * * Usage example: * * res = platform_get_resource(pdev, IORESOURCE_MEM, 0); * base = devm_ioremap_resource(&pdev->dev, res); * if (IS_ERR(base)) * return PTR_ERR(base); * * Return: a pointer to the remapped memory or an ERR_PTR() encoded error code * on failure.
*/ void __iomem *devm_ioremap_resource(struct device *dev, conststruct resource *res)
{ return __devm_ioremap_resource(dev, res, DEVM_IOREMAP);
}
EXPORT_SYMBOL(devm_ioremap_resource);
/** * devm_ioremap_resource_wc() - write-combined variant of * devm_ioremap_resource() * @dev: generic device to handle the resource for * @res: resource to be handled * * Return: a pointer to the remapped memory or an ERR_PTR() encoded error code * on failure.
*/ void __iomem *devm_ioremap_resource_wc(struct device *dev, conststruct resource *res)
{ return __devm_ioremap_resource(dev, res, DEVM_IOREMAP_WC);
}
EXPORT_SYMBOL(devm_ioremap_resource_wc);
/* * devm_of_iomap - Requests a resource and maps the memory mapped IO * for a given device_node managed by a given device * * Checks that a resource is a valid memory region, requests the memory * region and ioremaps it. All operations are managed and will be undone * on driver detach of the device. * * This is to be used when a device requests/maps resources described * by other device tree nodes (children or otherwise). * * @dev: The device "managing" the resource * @node: The device-tree node where the resource resides * @index: index of the MMIO range in the "reg" property * @size: Returns the size of the resource (pass NULL if not needed) * * Usage example: * * base = devm_of_iomap(&pdev->dev, node, 0, NULL); * if (IS_ERR(base)) * return PTR_ERR(base); * * Please Note: This is not a one-to-one replacement for of_iomap() because the * of_iomap() function does not track whether the region is already mapped. If * two drivers try to map the same memory, the of_iomap() function will succeed * but the devm_of_iomap() function will return -EBUSY. * * Return: a pointer to the requested and mapped memory or an ERR_PTR() encoded * error code on failure.
*/ void __iomem *devm_of_iomap(struct device *dev, struct device_node *node, int index,
resource_size_t *size)
{ struct resource res;
if (of_address_to_resource(node, index, &res)) return IOMEM_ERR_PTR(-EINVAL); if (size)
*size = resource_size(&res); return devm_ioremap_resource(dev, &res);
}
EXPORT_SYMBOL(devm_of_iomap);
/** * devm_ioport_map - Managed ioport_map() * @dev: Generic device to map ioport for * @port: Port to map * @nr: Number of ports to map * * Managed ioport_map(). Map is automatically unmapped on driver * detach. * * Return: a pointer to the remapped memory or NULL on failure.
*/ void __iomem *devm_ioport_map(struct device *dev, unsignedlong port, unsignedint nr)
{ void __iomem **ptr, *addr;
ptr = devres_alloc_node(devm_ioport_map_release, sizeof(*ptr), GFP_KERNEL,
dev_to_node(dev)); if (!ptr) return NULL;
/** * devm_arch_phys_wc_add - Managed arch_phys_wc_add() * @dev: Managed device * @base: Memory base address * @size: Size of memory range * * Adds a WC MTRR using arch_phys_wc_add() and sets up a release callback. * See arch_phys_wc_add() for more information.
*/ int devm_arch_phys_wc_add(struct device *dev, unsignedlong base, unsignedlong size)
{ int *mtrr; int ret;
mtrr = devres_alloc_node(devm_arch_phys_ac_add_release, sizeof(*mtrr), GFP_KERNEL,
dev_to_node(dev)); if (!mtrr) return -ENOMEM;
ret = arch_phys_wc_add(base, size); if (ret < 0) {
devres_free(mtrr); return ret;
}
/** * devm_arch_io_reserve_memtype_wc - Managed arch_io_reserve_memtype_wc() * @dev: Managed device * @start: Memory base address * @size: Size of memory range * * Reserves a memory range with WC caching using arch_io_reserve_memtype_wc() * and sets up a release callback See arch_io_reserve_memtype_wc() for more * information.
*/ int devm_arch_io_reserve_memtype_wc(struct device *dev, resource_size_t start,
resource_size_t size)
{ struct arch_io_reserve_memtype_wc_devres *dr; int ret;
dr = devres_alloc_node(devm_arch_io_free_memtype_wc_release, sizeof(*dr), GFP_KERNEL,
dev_to_node(dev)); if (!dr) return -ENOMEM;
ret = arch_io_reserve_memtype_wc(start, size); if (ret < 0) {
devres_free(dr); return ret;
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.