Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / gpu / drm / nouveau / nouveau_sgdma.c
diff --git a/kernel/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/kernel/drivers/gpu/drm/nouveau/nouveau_sgdma.c
new file mode 100644 (file)
index 0000000..8c3053a
--- /dev/null
@@ -0,0 +1,116 @@
+#include <linux/pagemap.h>
+#include <linux/slab.h>
+
+#include "nouveau_drm.h"
+#include "nouveau_ttm.h"
+
+struct nouveau_sgdma_be {
+       /* this has to be the first field so populate/unpopulated in
+        * nouve_bo.c works properly, otherwise have to move them here
+        */
+       struct ttm_dma_tt ttm;
+       struct nvkm_mem *node;
+};
+
+static void
+nouveau_sgdma_destroy(struct ttm_tt *ttm)
+{
+       struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
+
+       if (ttm) {
+               ttm_dma_tt_fini(&nvbe->ttm);
+               kfree(nvbe);
+       }
+}
+
+static int
+nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
+{
+       struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
+       struct nvkm_mem *node = mem->mm_node;
+
+       if (ttm->sg) {
+               node->sg    = ttm->sg;
+               node->pages = NULL;
+       } else {
+               node->sg    = NULL;
+               node->pages = nvbe->ttm.dma_address;
+       }
+       node->size = (mem->num_pages << PAGE_SHIFT) >> 12;
+
+       nvkm_vm_map(&node->vma[0], node);
+       nvbe->node = node;
+       return 0;
+}
+
+static int
+nv04_sgdma_unbind(struct ttm_tt *ttm)
+{
+       struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
+       nvkm_vm_unmap(&nvbe->node->vma[0]);
+       return 0;
+}
+
+static struct ttm_backend_func nv04_sgdma_backend = {
+       .bind                   = nv04_sgdma_bind,
+       .unbind                 = nv04_sgdma_unbind,
+       .destroy                = nouveau_sgdma_destroy
+};
+
+static int
+nv50_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem)
+{
+       struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm;
+       struct nvkm_mem *node = mem->mm_node;
+
+       /* noop: bound in move_notify() */
+       if (ttm->sg) {
+               node->sg    = ttm->sg;
+               node->pages = NULL;
+       } else {
+               node->sg    = NULL;
+               node->pages = nvbe->ttm.dma_address;
+       }
+       node->size = (mem->num_pages << PAGE_SHIFT) >> 12;
+       return 0;
+}
+
+static int
+nv50_sgdma_unbind(struct ttm_tt *ttm)
+{
+       /* noop: unbound in move_notify() */
+       return 0;
+}
+
+static struct ttm_backend_func nv50_sgdma_backend = {
+       .bind                   = nv50_sgdma_bind,
+       .unbind                 = nv50_sgdma_unbind,
+       .destroy                = nouveau_sgdma_destroy
+};
+
+struct ttm_tt *
+nouveau_sgdma_create_ttm(struct ttm_bo_device *bdev,
+                        unsigned long size, uint32_t page_flags,
+                        struct page *dummy_read_page)
+{
+       struct nouveau_drm *drm = nouveau_bdev(bdev);
+       struct nouveau_sgdma_be *nvbe;
+
+       nvbe = kzalloc(sizeof(*nvbe), GFP_KERNEL);
+       if (!nvbe)
+               return NULL;
+
+       if (drm->device.info.family < NV_DEVICE_INFO_V0_TESLA)
+               nvbe->ttm.ttm.func = &nv04_sgdma_backend;
+       else
+               nvbe->ttm.ttm.func = &nv50_sgdma_backend;
+
+       if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page))
+               /*
+                * A failing ttm_dma_tt_init() will call ttm_tt_destroy()
+                * and thus our nouveau_sgdma_destroy() hook, so we don't need
+                * to free nvbe here.
+                */
+               return NULL;
+       return &nvbe->ttm.ttm;
+}