Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / gpu / drm / exynos / exynos_drm_buf.c
1 /* exynos_drm_buf.c
2  *
3  * Copyright (c) 2011 Samsung Electronics Co., Ltd.
4  * Author: Inki Dae <inki.dae@samsung.com>
5  *
6  * This program is free software; you can redistribute  it and/or modify it
7  * under  the terms of  the GNU General  Public License as published by the
8  * Free Software Foundation;  either version 2 of the  License, or (at your
9  * option) any later version.
10  */
11
12 #include <drm/drmP.h>
13 #include <drm/exynos_drm.h>
14
15 #include "exynos_drm_drv.h"
16 #include "exynos_drm_gem.h"
17 #include "exynos_drm_buf.h"
18 #include "exynos_drm_iommu.h"
19
20 static int lowlevel_buffer_allocate(struct drm_device *dev,
21                 unsigned int flags, struct exynos_drm_gem_buf *buf)
22 {
23         int ret = 0;
24         enum dma_attr attr;
25         unsigned int nr_pages;
26
27         if (buf->dma_addr) {
28                 DRM_DEBUG_KMS("already allocated.\n");
29                 return 0;
30         }
31
32         init_dma_attrs(&buf->dma_attrs);
33
34         /*
35          * if EXYNOS_BO_CONTIG, fully physically contiguous memory
36          * region will be allocated else physically contiguous
37          * as possible.
38          */
39         if (!(flags & EXYNOS_BO_NONCONTIG))
40                 dma_set_attr(DMA_ATTR_FORCE_CONTIGUOUS, &buf->dma_attrs);
41
42         /*
43          * if EXYNOS_BO_WC or EXYNOS_BO_NONCACHABLE, writecombine mapping
44          * else cachable mapping.
45          */
46         if (flags & EXYNOS_BO_WC || !(flags & EXYNOS_BO_CACHABLE))
47                 attr = DMA_ATTR_WRITE_COMBINE;
48         else
49                 attr = DMA_ATTR_NON_CONSISTENT;
50
51         dma_set_attr(attr, &buf->dma_attrs);
52         dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &buf->dma_attrs);
53
54         nr_pages = buf->size >> PAGE_SHIFT;
55
56         if (!is_drm_iommu_supported(dev)) {
57                 dma_addr_t start_addr;
58                 unsigned int i = 0;
59
60                 buf->pages = drm_calloc_large(nr_pages, sizeof(struct page *));
61                 if (!buf->pages) {
62                         DRM_ERROR("failed to allocate pages.\n");
63                         return -ENOMEM;
64                 }
65
66                 buf->cookie = dma_alloc_attrs(dev->dev,
67                                         buf->size,
68                                         &buf->dma_addr, GFP_KERNEL,
69                                         &buf->dma_attrs);
70                 if (!buf->cookie) {
71                         DRM_ERROR("failed to allocate buffer.\n");
72                         ret = -ENOMEM;
73                         goto err_free;
74                 }
75
76                 start_addr = buf->dma_addr;
77                 while (i < nr_pages) {
78                         buf->pages[i] = phys_to_page(start_addr);
79                         start_addr += PAGE_SIZE;
80                         i++;
81                 }
82         } else {
83
84                 buf->pages = dma_alloc_attrs(dev->dev, buf->size,
85                                         &buf->dma_addr, GFP_KERNEL,
86                                         &buf->dma_attrs);
87                 if (!buf->pages) {
88                         DRM_ERROR("failed to allocate buffer.\n");
89                         return -ENOMEM;
90                 }
91         }
92
93         buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages);
94         if (IS_ERR(buf->sgt)) {
95                 DRM_ERROR("failed to get sg table.\n");
96                 ret = PTR_ERR(buf->sgt);
97                 goto err_free_attrs;
98         }
99
100         DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
101                         (unsigned long)buf->dma_addr,
102                         buf->size);
103
104         return ret;
105
106 err_free_attrs:
107         dma_free_attrs(dev->dev, buf->size, buf->pages,
108                         (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
109         buf->dma_addr = (dma_addr_t)NULL;
110 err_free:
111         if (!is_drm_iommu_supported(dev))
112                 drm_free_large(buf->pages);
113
114         return ret;
115 }
116
117 static void lowlevel_buffer_deallocate(struct drm_device *dev,
118                 unsigned int flags, struct exynos_drm_gem_buf *buf)
119 {
120         if (!buf->dma_addr) {
121                 DRM_DEBUG_KMS("dma_addr is invalid.\n");
122                 return;
123         }
124
125         DRM_DEBUG_KMS("dma_addr(0x%lx), size(0x%lx)\n",
126                         (unsigned long)buf->dma_addr,
127                         buf->size);
128
129         sg_free_table(buf->sgt);
130
131         kfree(buf->sgt);
132         buf->sgt = NULL;
133
134         if (!is_drm_iommu_supported(dev)) {
135                 dma_free_attrs(dev->dev, buf->size, buf->cookie,
136                                 (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
137                 drm_free_large(buf->pages);
138         } else
139                 dma_free_attrs(dev->dev, buf->size, buf->pages,
140                                 (dma_addr_t)buf->dma_addr, &buf->dma_attrs);
141
142         buf->dma_addr = (dma_addr_t)NULL;
143 }
144
145 struct exynos_drm_gem_buf *exynos_drm_init_buf(struct drm_device *dev,
146                                                 unsigned int size)
147 {
148         struct exynos_drm_gem_buf *buffer;
149
150         DRM_DEBUG_KMS("desired size = 0x%x\n", size);
151
152         buffer = kzalloc(sizeof(*buffer), GFP_KERNEL);
153         if (!buffer)
154                 return NULL;
155
156         buffer->size = size;
157         return buffer;
158 }
159
160 void exynos_drm_fini_buf(struct drm_device *dev,
161                                 struct exynos_drm_gem_buf *buffer)
162 {
163         kfree(buffer);
164         buffer = NULL;
165 }
166
167 int exynos_drm_alloc_buf(struct drm_device *dev,
168                 struct exynos_drm_gem_buf *buf, unsigned int flags)
169 {
170
171         /*
172          * allocate memory region and set the memory information
173          * to vaddr and dma_addr of a buffer object.
174          */
175         if (lowlevel_buffer_allocate(dev, flags, buf) < 0)
176                 return -ENOMEM;
177
178         return 0;
179 }
180
181 void exynos_drm_free_buf(struct drm_device *dev,
182                 unsigned int flags, struct exynos_drm_gem_buf *buffer)
183 {
184
185         lowlevel_buffer_deallocate(dev, flags, buffer);
186 }