Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / gpu / drm / exynos / exynos_drm_buf.c
diff --git a/kernel/drivers/gpu/drm/exynos/exynos_drm_buf.c b/kernel/drivers/gpu/drm/exynos/exynos_drm_buf.c
new file mode 100644 (file)
index 0000000..24994ba
--- /dev/null
@@ -0,0 +1,186 @@
+/* exynos_drm_buf.c
+ *
+ * Copyright (c) 2011 Samsung Electronics Co., Ltd.
+ * Author: Inki Dae <inki.dae@samsung.com>
+ *
+ * This program is free software; you can redistribute  it and/or modify it
+ * under  the terms of  the GNU General  Public License as published by the
+ * Free Software Foundation;  either version 2 of the  License, or (at your
+ * option) any later version.
+ */
+
+#include <drm/drmP.h>
+#include <drm/exynos_drm.h>
+
+#include "exynos_drm_drv.h"
+#include "exynos_drm_gem.h"
+#include "exynos_drm_buf.h"
+#include "exynos_drm_iommu.h"
+
+static int lowlevel_buffer_allocate(struct drm_device *dev,
+               unsigned int flags, struct exynos_drm_gem_buf *buf)
+{
+       int ret = 0;
+       enum dma_attr attr;
+       unsigned int nr_pages;
+
+       if (buf->dma_addr) {
+               DRM_DEBUG_KMS("already allocated.\n");
+               return 0;
+       }
+
+       init_dma_attrs(&buf->dma_attrs);
+
+       /*
+        * if EXYNOS_BO_CONTIG, fully physically contiguous memory
+        * region will be allocated else physically contiguous
+        * as possible.
+        */
+       if (!(flags & EXYNOS_BO_NONCONTIG))
+               dma_set_attr(DMA_ATTR_FORCE_CONTIGUOUS, &buf->dma_attrs);
+
+       /*
+        * if EXYNOS_BO_WC or EXYNOS_BO_NONCACHABLE, writecombine mapping
+        * else cachable mapping.
+        */
+       if (flags & EXYNOS_BO_WC || !(flags & EXYNOS_BO_CACHABLE))
+               attr = DMA_ATTR_WRITE_COMBINE;
+       else
+               attr = DMA_ATTR_NON_CONSISTENT;
+
+       dma_set_attr(attr, &buf->dma_attrs);
+       dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &buf->dma_attrs);
+
+       nr_pages = buf->size >> PAGE_SHIFT;
+
+       if (!is_drm_iommu_supported(dev)) {
+               dma_addr_t start_addr;
+               unsigned int i = 0;
+
+               buf->pages = drm_calloc_large(nr_pages, sizeof(struct page *));
+               if (!buf->pages) {
+                       DRM_ERROR("failed to allocate pages.\n");
+                       return -ENOMEM;
+               }
+
+               buf->cookie = dma_alloc_attrs(dev->dev,
+                                       buf->size,
+                                       &buf->dma_addr, GFP_KERNEL,
+                                       &buf->dma_attrs);
+               if (!buf->cookie) {
+                       DRM_ERROR("failed to allocate buffer.\n");
+                       ret = -ENOMEM;
+                       goto err_free;
+               }
+
+               start_addr = buf->dma_addr;
+               while (i < nr_pages) {
+                       buf->pages[i] = phys_to_page(start_addr);
+                       start_addr += PAGE_SIZE;
+                       i++;
+               }
+       } else {
+
+               buf->pages = dma_alloc_attrs(dev->dev, buf->size,
+                                       &buf->dma_addr, GFP_KERNEL,
+                                       &buf->dma_attrs);
+               if (!buf->pages) {
+                       DRM_ERROR("failed to allocate buffer.\n");
+                       return -ENOMEM;
+               }
+       }
+
+       buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages);
+       if (IS_ERR(buf->sgt)) {
+               DRM_ERROR("failed to get sg table.\n");
+               ret = PTR_ERR(buf->sgt);
+               goto err_free_attrs;
+       }
+
+       DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
+                       (unsigned long)buf->dma_addr,
+                       buf->size);
+
+       return ret;
+
+err_free_attrs:
+       dma_free_attrs(dev->dev, buf->size, buf->pages,
+                       (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
+       buf->dma_addr = (dma_addr_t)NULL;
+err_free:
+       if (!is_drm_iommu_supported(dev))
+               drm_free_large(buf->pages);
+
+       return ret;
+}
+
+static void lowlevel_buffer_deallocate(struct drm_device *dev,
+               unsigned int flags, struct exynos_drm_gem_buf *buf)
+{
+       if (!buf->dma_addr) {
+               DRM_DEBUG_KMS("dma_addr is invalid.\n");
+               return;
+       }
+
+       DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
+                       (unsigned long)buf->dma_addr,
+                       buf->size);
+
+       sg_free_table(buf->sgt);
+
+       kfree(buf->sgt);
+       buf->sgt = NULL;
+
+       if (!is_drm_iommu_supported(dev)) {
+               dma_free_attrs(dev->dev, buf->size, buf->cookie,
+                               (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
+               drm_free_large(buf->pages);
+       } else
+               dma_free_attrs(dev->dev, buf->size, buf->pages,
+                               (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
+
+       buf->dma_addr = (dma_addr_t)NULL;
+}
+
+struct exynos_drm_gem_buf *exynos_drm_init_buf(struct drm_device *dev,
+                                               unsigned int size)
+{
+       struct exynos_drm_gem_buf *buffer;
+
+       DRM_DEBUG_KMS("desired size = 0x%x\n", size);
+
+       buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
+       if (!buffer)
+               return NULL;
+
+       buffer->size = size;
+       return buffer;
+}
+
+void exynos_drm_fini_buf(struct drm_device *dev,
+                               struct exynos_drm_gem_buf *buffer)
+{
+       kfree(buffer);
+       buffer = NULL;
+}
+
+int exynos_drm_alloc_buf(struct drm_device *dev,
+               struct exynos_drm_gem_buf *buf, unsigned int flags)
+{
+
+       /*
+        * allocate memory region and set the memory information
+        * to vaddr and dma_addr of a buffer object.
+        */
+       if (lowlevel_buffer_allocate(dev, flags, buf) < 0)
+               return -ENOMEM;
+
+       return 0;
+}
+
+void exynos_drm_free_buf(struct drm_device *dev,
+               unsigned int flags, struct exynos_drm_gem_buf *buffer)
+{
+
+       lowlevel_buffer_deallocate(dev, flags, buffer);
+}