/* * We are not syncing against the binding (and potential migrations) * below, so this vm must never be async.
*/ if (drm_WARN_ON(&dev_priv->drm, vm->bind_async_flags)) return ERR_PTR(-EINVAL);
if (WARN_ON(!i915_gem_object_is_framebuffer(obj))) return ERR_PTR(-EINVAL);
atomic_inc(&display->restore.pending_fb_pin);
for_i915_gem_ww(&ww, ret, true) {
ret = i915_gem_object_lock(obj, &ww); if (ret) continue;
if (HAS_LMEM(dev_priv)) { unsignedint flags = obj->flags;
/* * For this type of buffer we need to able to read from the CPU * the clear color value found in the buffer, hence we need to * ensure it is always in the mappable part of lmem, if this is * a small-bar device.
*/ if (intel_fb_rc_ccs_cc_plane(fb) >= 0)
flags &= ~I915_BO_ALLOC_GPU_ONLY;
ret = __i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0,
flags); if (ret) continue;
}
ret = i915_gem_object_set_cache_level(obj, I915_CACHE_NONE); if (ret) continue;
vma = i915_vma_instance(obj, vm, view); if (IS_ERR(vma)) {
ret = PTR_ERR(vma); continue;
}
if (i915_vma_misplaced(vma, 0, alignment, 0)) {
ret = i915_vma_unbind(vma); if (ret) continue;
}
ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL); if (ret) continue;
} if (ret) {
vma = ERR_PTR(ret); goto err;
}
if (drm_WARN_ON(dev, !i915_gem_object_is_framebuffer(obj))) return ERR_PTR(-EINVAL);
if (drm_WARN_ON(dev, alignment && !is_power_of_2(alignment))) return ERR_PTR(-EINVAL);
/* * Global gtt pte registers are special registers which actually forward * writes to a chunk of system memory. Which means that there is no risk * that the register values disappear as soon as we call * intel_runtime_pm_put(), so it is correct to wrap only the * pin/unpin/fence and not more.
*/
wakeref = intel_display_rpm_get(display);
atomic_inc(&display->restore.pending_fb_pin);
/* * Valleyview is definitely limited to scanning out the first * 512MiB. Lets presume this behaviour was inherited from the * g4x display engine and that all earlier gen are similarly * limited. Testing suggests that it is a little more * complicated than this. For example, Cherryview appears quite * happy to scanout from anywhere within its global aperture.
*/
pinctl = 0; if (HAS_GMCH(dev_priv))
pinctl |= PIN_MAPPABLE;
i915_gem_ww_ctx_init(&ww, true);
retry:
ret = i915_gem_object_lock(obj, &ww); if (!ret && phys_alignment)
ret = i915_gem_object_attach_phys(obj, phys_alignment); elseif (!ret && HAS_LMEM(dev_priv))
ret = i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0); if (!ret)
ret = i915_gem_object_pin_pages(obj); if (ret) goto err;
vma = i915_gem_object_pin_to_display_plane(obj, &ww, alignment,
vtd_guard, view, pinctl); if (IS_ERR(vma)) {
ret = PTR_ERR(vma); goto err_unpin;
}
if (uses_fence && i915_vma_is_map_and_fenceable(vma)) { /* * Install a fence for tiled scan-out. Pre-i965 always needs a * fence, whereas 965+ only requires a fence if using * framebuffer compression. For simplicity, we always, when * possible, install a fence as the cost is not that onerous. * * If we fail to fence the tiled scanout, then either the * modeset will reject the change (which is highly unlikely as * the affected systems, all but one, do not have unmappable * space) or we will not be able to enable full powersaving * techniques (also likely not to apply due to various limits * FBC and the like impose on the size of the buffer, which * presumably we violated anyway with this unmappable buffer). * Anyway, it is presumably better to stumble onwards with * something and try to run the system in a "less than optimal" * mode that matches the user configuration.
*/
ret = i915_vma_pin_fence(vma); if (ret != 0 && DISPLAY_VER(dev_priv) < 4) {
i915_vma_unpin(vma); goto err_unpin;
}
ret = 0;
if (vma->fence)
*out_flags |= PLANE_HAS_FENCE;
}
i915_vma_get(vma);
err_unpin:
i915_gem_object_unpin_pages(obj);
err: if (ret == -EDEADLK) {
ret = i915_gem_ww_ctx_backoff(&ww); if (!ret) goto retry;
}
i915_gem_ww_ctx_fini(&ww); if (ret)
vma = ERR_PTR(ret);
if (!intel_fb_uses_dpt(&fb->base)) {
vma = intel_fb_pin_to_ggtt(&fb->base, &plane_state->view.gtt,
intel_plane_fb_min_alignment(plane_state),
intel_plane_fb_min_phys_alignment(plane_state),
intel_plane_fb_vtd_guard(plane_state),
intel_plane_uses_fence(plane_state),
&plane_state->flags); if (IS_ERR(vma)) return PTR_ERR(vma);
plane_state->ggtt_vma = vma;
/* * Pre-populate the dma address before we enter the vblank * evade critical section as i915_gem_object_get_dma_address() * will trigger might_sleep() even if it won't actually sleep, * which is the case when the fb has already been pinned.
*/ if (intel_plane_needs_physical(plane)) { struct drm_i915_gem_object *obj = to_intel_bo(intel_fb_bo(&fb->base));
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.