i915_gem_gtt.c 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157
  1. /*
  2. * Copyright © 2010 Daniel Vetter
  3. *
  4. * Permission is hereby granted, free of charge, to any person obtaining a
  5. * copy of this software and associated documentation files (the "Software"),
  6. * to deal in the Software without restriction, including without limitation
  7. * the rights to use, copy, modify, merge, publish, distribute, sublicense,
  8. * and/or sell copies of the Software, and to permit persons to whom the
  9. * Software is furnished to do so, subject to the following conditions:
  10. *
  11. * The above copyright notice and this permission notice (including the next
  12. * paragraph) shall be included in all copies or substantial portions of the
  13. * Software.
  14. *
  15. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  16. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  17. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  18. * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  19. * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
  20. * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
  21. * IN THE SOFTWARE.
  22. *
  23. */
  24. #include "drmP.h"
  25. #include "drm.h"
  26. #include "i915_drm.h"
  27. #include "i915_drv.h"
  28. #include "i915_trace.h"
  29. #include "intel_drv.h"
  30. /* XXX kill agp_type! */
  31. static unsigned int cache_level_to_agp_type(struct drm_device *dev,
  32. enum i915_cache_level cache_level)
  33. {
  34. switch (cache_level) {
  35. case I915_CACHE_LLC_MLC:
  36. if (INTEL_INFO(dev)->gen >= 6)
  37. return AGP_USER_CACHED_MEMORY_LLC_MLC;
  38. /* Older chipsets do not have this extra level of CPU
  39. * cacheing, so fallthrough and request the PTE simply
  40. * as cached.
  41. */
  42. case I915_CACHE_LLC:
  43. return AGP_USER_CACHED_MEMORY;
  44. default:
  45. case I915_CACHE_NONE:
  46. return AGP_USER_MEMORY;
  47. }
  48. }
  49. static bool do_idling(struct drm_i915_private *dev_priv)
  50. {
  51. bool ret = dev_priv->mm.interruptible;
  52. if (unlikely(dev_priv->mm.gtt->do_idle_maps)) {
  53. dev_priv->mm.interruptible = false;
  54. if (i915_gpu_idle(dev_priv->dev)) {
  55. DRM_ERROR("Couldn't idle GPU\n");
  56. /* Wait a bit, in hopes it avoids the hang */
  57. udelay(10);
  58. }
  59. }
  60. return ret;
  61. }
  62. static void undo_idling(struct drm_i915_private *dev_priv, bool interruptible)
  63. {
  64. if (unlikely(dev_priv->mm.gtt->do_idle_maps))
  65. dev_priv->mm.interruptible = interruptible;
  66. }
  67. void i915_gem_restore_gtt_mappings(struct drm_device *dev)
  68. {
  69. struct drm_i915_private *dev_priv = dev->dev_private;
  70. struct drm_i915_gem_object *obj;
  71. /* First fill our portion of the GTT with scratch pages */
  72. intel_gtt_clear_range(dev_priv->mm.gtt_start / PAGE_SIZE,
  73. (dev_priv->mm.gtt_end - dev_priv->mm.gtt_start) / PAGE_SIZE);
  74. list_for_each_entry(obj, &dev_priv->mm.gtt_list, gtt_list) {
  75. i915_gem_clflush_object(obj);
  76. i915_gem_gtt_rebind_object(obj, obj->cache_level);
  77. }
  78. intel_gtt_chipset_flush();
  79. }
  80. int i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj)
  81. {
  82. struct drm_device *dev = obj->base.dev;
  83. struct drm_i915_private *dev_priv = dev->dev_private;
  84. unsigned int agp_type = cache_level_to_agp_type(dev, obj->cache_level);
  85. int ret;
  86. if (dev_priv->mm.gtt->needs_dmar) {
  87. ret = intel_gtt_map_memory(obj->pages,
  88. obj->base.size >> PAGE_SHIFT,
  89. &obj->sg_list,
  90. &obj->num_sg);
  91. if (ret != 0)
  92. return ret;
  93. intel_gtt_insert_sg_entries(obj->sg_list,
  94. obj->num_sg,
  95. obj->gtt_space->start >> PAGE_SHIFT,
  96. agp_type);
  97. } else
  98. intel_gtt_insert_pages(obj->gtt_space->start >> PAGE_SHIFT,
  99. obj->base.size >> PAGE_SHIFT,
  100. obj->pages,
  101. agp_type);
  102. return 0;
  103. }
  104. void i915_gem_gtt_rebind_object(struct drm_i915_gem_object *obj,
  105. enum i915_cache_level cache_level)
  106. {
  107. struct drm_device *dev = obj->base.dev;
  108. struct drm_i915_private *dev_priv = dev->dev_private;
  109. unsigned int agp_type = cache_level_to_agp_type(dev, cache_level);
  110. if (dev_priv->mm.gtt->needs_dmar) {
  111. BUG_ON(!obj->sg_list);
  112. intel_gtt_insert_sg_entries(obj->sg_list,
  113. obj->num_sg,
  114. obj->gtt_space->start >> PAGE_SHIFT,
  115. agp_type);
  116. } else
  117. intel_gtt_insert_pages(obj->gtt_space->start >> PAGE_SHIFT,
  118. obj->base.size >> PAGE_SHIFT,
  119. obj->pages,
  120. agp_type);
  121. }
  122. void i915_gem_gtt_unbind_object(struct drm_i915_gem_object *obj)
  123. {
  124. struct drm_device *dev = obj->base.dev;
  125. struct drm_i915_private *dev_priv = dev->dev_private;
  126. bool interruptible;
  127. interruptible = do_idling(dev_priv);
  128. intel_gtt_clear_range(obj->gtt_space->start >> PAGE_SHIFT,
  129. obj->base.size >> PAGE_SHIFT);
  130. if (obj->sg_list) {
  131. intel_gtt_unmap_memory(obj->sg_list, obj->num_sg);
  132. obj->sg_list = NULL;
  133. }
  134. undo_idling(dev_priv, interruptible);
  135. }