Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / gpu / drm / sti / sti_gdp.c
diff --git a/kernel/drivers/gpu/drm/sti/sti_gdp.c b/kernel/drivers/gpu/drm/sti/sti_gdp.c
new file mode 100644 (file)
index 0000000..087906f
--- /dev/null
@@ -0,0 +1,584 @@
+/*
+ * Copyright (C) STMicroelectronics SA 2014
+ * Authors: Benjamin Gaignard <benjamin.gaignard@st.com>
+ *          Fabien Dessenne <fabien.dessenne@st.com>
+ *          for STMicroelectronics.
+ * License terms:  GNU General Public License (GPL), version 2
+ */
+
+#include <linux/clk.h>
+#include <linux/dma-mapping.h>
+
+#include "sti_compositor.h"
+#include "sti_gdp.h"
+#include "sti_layer.h"
+#include "sti_vtg.h"
+
+#define ALPHASWITCH     BIT(6)
+#define ENA_COLOR_FILL  BIT(8)
+#define BIGNOTLITTLE    BIT(23)
+#define WAIT_NEXT_VSYNC BIT(31)
+
+/* GDP color formats */
+#define GDP_RGB565      0x00
+#define GDP_RGB888      0x01
+#define GDP_RGB888_32   0x02
+#define GDP_XBGR8888    (GDP_RGB888_32 | BIGNOTLITTLE | ALPHASWITCH)
+#define GDP_ARGB8565    0x04
+#define GDP_ARGB8888    0x05
+#define GDP_ABGR8888   (GDP_ARGB8888 | BIGNOTLITTLE | ALPHASWITCH)
+#define GDP_ARGB1555    0x06
+#define GDP_ARGB4444    0x07
+#define GDP_CLUT8       0x0B
+#define GDP_YCBR888     0x10
+#define GDP_YCBR422R    0x12
+#define GDP_AYCBR8888   0x15
+
+#define GAM_GDP_CTL_OFFSET      0x00
+#define GAM_GDP_AGC_OFFSET      0x04
+#define GAM_GDP_VPO_OFFSET      0x0C
+#define GAM_GDP_VPS_OFFSET      0x10
+#define GAM_GDP_PML_OFFSET      0x14
+#define GAM_GDP_PMP_OFFSET      0x18
+#define GAM_GDP_SIZE_OFFSET     0x1C
+#define GAM_GDP_NVN_OFFSET      0x24
+#define GAM_GDP_KEY1_OFFSET     0x28
+#define GAM_GDP_KEY2_OFFSET     0x2C
+#define GAM_GDP_PPT_OFFSET      0x34
+#define GAM_GDP_CML_OFFSET      0x3C
+#define GAM_GDP_MST_OFFSET      0x68
+
+#define GAM_GDP_ALPHARANGE_255  BIT(5)
+#define GAM_GDP_AGC_FULL_RANGE  0x00808080
+#define GAM_GDP_PPT_IGNORE      (BIT(1) | BIT(0))
+#define GAM_GDP_SIZE_MAX        0x7FF
+
+#define GDP_NODE_NB_BANK       2
+#define GDP_NODE_PER_FIELD     2
+
+struct sti_gdp_node {
+       u32 gam_gdp_ctl;
+       u32 gam_gdp_agc;
+       u32 reserved1;
+       u32 gam_gdp_vpo;
+       u32 gam_gdp_vps;
+       u32 gam_gdp_pml;
+       u32 gam_gdp_pmp;
+       u32 gam_gdp_size;
+       u32 reserved2;
+       u32 gam_gdp_nvn;
+       u32 gam_gdp_key1;
+       u32 gam_gdp_key2;
+       u32 reserved3;
+       u32 gam_gdp_ppt;
+       u32 reserved4;
+       u32 gam_gdp_cml;
+};
+
+struct sti_gdp_node_list {
+       struct sti_gdp_node *top_field;
+       dma_addr_t top_field_paddr;
+       struct sti_gdp_node *btm_field;
+       dma_addr_t btm_field_paddr;
+};
+
+/**
+ * STI GDP structure
+ *
+ * @layer:             layer structure
+ * @clk_pix:            pixel clock for the current gdp
+ * @clk_main_parent:    gdp parent clock if main path used
+ * @clk_aux_parent:     gdp parent clock if aux path used
+ * @vtg_field_nb:       callback for VTG FIELD (top or bottom) notification
+ * @is_curr_top:        true if the current node processed is the top field
+ * @node_list:         array of node list
+ */
+struct sti_gdp {
+       struct sti_layer layer;
+       struct clk *clk_pix;
+       struct clk *clk_main_parent;
+       struct clk *clk_aux_parent;
+       struct notifier_block vtg_field_nb;
+       bool is_curr_top;
+       struct sti_gdp_node_list node_list[GDP_NODE_NB_BANK];
+};
+
+#define to_sti_gdp(x) container_of(x, struct sti_gdp, layer)
+
+static const uint32_t gdp_supported_formats[] = {
+       DRM_FORMAT_XRGB8888,
+       DRM_FORMAT_XBGR8888,
+       DRM_FORMAT_ARGB8888,
+       DRM_FORMAT_ABGR8888,
+       DRM_FORMAT_ARGB4444,
+       DRM_FORMAT_ARGB1555,
+       DRM_FORMAT_RGB565,
+       DRM_FORMAT_RGB888,
+       DRM_FORMAT_AYUV,
+       DRM_FORMAT_YUV444,
+       DRM_FORMAT_VYUY,
+       DRM_FORMAT_C8,
+};
+
+static const uint32_t *sti_gdp_get_formats(struct sti_layer *layer)
+{
+       return gdp_supported_formats;
+}
+
+static unsigned int sti_gdp_get_nb_formats(struct sti_layer *layer)
+{
+       return ARRAY_SIZE(gdp_supported_formats);
+}
+
+static int sti_gdp_fourcc2format(int fourcc)
+{
+       switch (fourcc) {
+       case DRM_FORMAT_XRGB8888:
+               return GDP_RGB888_32;
+       case DRM_FORMAT_XBGR8888:
+               return GDP_XBGR8888;
+       case DRM_FORMAT_ARGB8888:
+               return GDP_ARGB8888;
+       case DRM_FORMAT_ABGR8888:
+               return GDP_ABGR8888;
+       case DRM_FORMAT_ARGB4444:
+               return GDP_ARGB4444;
+       case DRM_FORMAT_ARGB1555:
+               return GDP_ARGB1555;
+       case DRM_FORMAT_RGB565:
+               return GDP_RGB565;
+       case DRM_FORMAT_RGB888:
+               return GDP_RGB888;
+       case DRM_FORMAT_AYUV:
+               return GDP_AYCBR8888;
+       case DRM_FORMAT_YUV444:
+               return GDP_YCBR888;
+       case DRM_FORMAT_VYUY:
+               return GDP_YCBR422R;
+       case DRM_FORMAT_C8:
+               return GDP_CLUT8;
+       }
+       return -1;
+}
+
+static int sti_gdp_get_alpharange(int format)
+{
+       switch (format) {
+       case GDP_ARGB8565:
+       case GDP_ARGB8888:
+       case GDP_AYCBR8888:
+       case GDP_ABGR8888:
+               return GAM_GDP_ALPHARANGE_255;
+       }
+       return 0;
+}
+
+/**
+ * sti_gdp_get_free_nodes
+ * @layer: gdp layer
+ *
+ * Look for a GDP node list that is not currently read by the HW.
+ *
+ * RETURNS:
+ * Pointer to the free GDP node list
+ */
+static struct sti_gdp_node_list *sti_gdp_get_free_nodes(struct sti_layer *layer)
+{
+       int hw_nvn;
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       unsigned int i;
+
+       hw_nvn = readl(layer->regs + GAM_GDP_NVN_OFFSET);
+       if (!hw_nvn)
+               goto end;
+
+       for (i = 0; i < GDP_NODE_NB_BANK; i++)
+               if ((hw_nvn != gdp->node_list[i].btm_field_paddr) &&
+                   (hw_nvn != gdp->node_list[i].top_field_paddr))
+                       return &gdp->node_list[i];
+
+       /* in hazardious cases restart with the first node */
+       DRM_ERROR("inconsistent NVN for %s: 0x%08X\n",
+                       sti_layer_to_str(layer), hw_nvn);
+
+end:
+       return &gdp->node_list[0];
+}
+
+/**
+ * sti_gdp_get_current_nodes
+ * @layer: GDP layer
+ *
+ * Look for GDP nodes that are currently read by the HW.
+ *
+ * RETURNS:
+ * Pointer to the current GDP node list
+ */
+static
+struct sti_gdp_node_list *sti_gdp_get_current_nodes(struct sti_layer *layer)
+{
+       int hw_nvn;
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       unsigned int i;
+
+       hw_nvn = readl(layer->regs + GAM_GDP_NVN_OFFSET);
+       if (!hw_nvn)
+               goto end;
+
+       for (i = 0; i < GDP_NODE_NB_BANK; i++)
+               if ((hw_nvn == gdp->node_list[i].btm_field_paddr) ||
+                               (hw_nvn == gdp->node_list[i].top_field_paddr))
+                       return &gdp->node_list[i];
+
+end:
+       DRM_DEBUG_DRIVER("Warning, NVN 0x%08X for %s does not match any node\n",
+                               hw_nvn, sti_layer_to_str(layer));
+
+       return NULL;
+}
+
+/**
+ * sti_gdp_prepare_layer
+ * @lay: gdp layer
+ * @first_prepare: true if it is the first time this function is called
+ *
+ * Update the free GDP node list according to the layer properties.
+ *
+ * RETURNS:
+ * 0 on success.
+ */
+static int sti_gdp_prepare_layer(struct sti_layer *layer, bool first_prepare)
+{
+       struct sti_gdp_node_list *list;
+       struct sti_gdp_node *top_field, *btm_field;
+       struct drm_display_mode *mode = layer->mode;
+       struct device *dev = layer->dev;
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       struct sti_compositor *compo = dev_get_drvdata(dev);
+       int format;
+       unsigned int depth, bpp;
+       int rate = mode->clock * 1000;
+       int res;
+       u32 ydo, xdo, yds, xds;
+
+       list = sti_gdp_get_free_nodes(layer);
+       top_field = list->top_field;
+       btm_field = list->btm_field;
+
+       dev_dbg(dev, "%s %s top_node:0x%p btm_node:0x%p\n", __func__,
+                       sti_layer_to_str(layer), top_field, btm_field);
+
+       /* Build the top field from layer params */
+       top_field->gam_gdp_agc = GAM_GDP_AGC_FULL_RANGE;
+       top_field->gam_gdp_ctl = WAIT_NEXT_VSYNC;
+       format = sti_gdp_fourcc2format(layer->format);
+       if (format == -1) {
+               DRM_ERROR("Format not supported by GDP %.4s\n",
+                         (char *)&layer->format);
+               return 1;
+       }
+       top_field->gam_gdp_ctl |= format;
+       top_field->gam_gdp_ctl |= sti_gdp_get_alpharange(format);
+       top_field->gam_gdp_ppt &= ~GAM_GDP_PPT_IGNORE;
+
+       /* pixel memory location */
+       drm_fb_get_bpp_depth(layer->format, &depth, &bpp);
+       top_field->gam_gdp_pml = (u32) layer->paddr + layer->offsets[0];
+       top_field->gam_gdp_pml += layer->src_x * (bpp >> 3);
+       top_field->gam_gdp_pml += layer->src_y * layer->pitches[0];
+
+       /* input parameters */
+       top_field->gam_gdp_pmp = layer->pitches[0];
+       top_field->gam_gdp_size =
+           clamp_val(layer->src_h, 0, GAM_GDP_SIZE_MAX) << 16 |
+           clamp_val(layer->src_w, 0, GAM_GDP_SIZE_MAX);
+
+       /* output parameters */
+       ydo = sti_vtg_get_line_number(*mode, layer->dst_y);
+       yds = sti_vtg_get_line_number(*mode, layer->dst_y + layer->dst_h - 1);
+       xdo = sti_vtg_get_pixel_number(*mode, layer->dst_x);
+       xds = sti_vtg_get_pixel_number(*mode, layer->dst_x + layer->dst_w - 1);
+       top_field->gam_gdp_vpo = (ydo << 16) | xdo;
+       top_field->gam_gdp_vps = (yds << 16) | xds;
+
+       /* Same content and chained together */
+       memcpy(btm_field, top_field, sizeof(*btm_field));
+       top_field->gam_gdp_nvn = list->btm_field_paddr;
+       btm_field->gam_gdp_nvn = list->top_field_paddr;
+
+       /* Interlaced mode */
+       if (layer->mode->flags & DRM_MODE_FLAG_INTERLACE)
+               btm_field->gam_gdp_pml = top_field->gam_gdp_pml +
+                   layer->pitches[0];
+
+       if (first_prepare) {
+               /* Register gdp callback */
+               if (sti_vtg_register_client(layer->mixer_id == STI_MIXER_MAIN ?
+                               compo->vtg_main : compo->vtg_aux,
+                               &gdp->vtg_field_nb, layer->mixer_id)) {
+                       DRM_ERROR("Cannot register VTG notifier\n");
+                       return 1;
+               }
+
+               /* Set and enable gdp clock */
+               if (gdp->clk_pix) {
+                       struct clk *clkp;
+                       /* According to the mixer used, the gdp pixel clock
+                        * should have a different parent clock. */
+                       if (layer->mixer_id == STI_MIXER_MAIN)
+                               clkp = gdp->clk_main_parent;
+                       else
+                               clkp = gdp->clk_aux_parent;
+
+                       if (clkp)
+                               clk_set_parent(gdp->clk_pix, clkp);
+
+                       res = clk_set_rate(gdp->clk_pix, rate);
+                       if (res < 0) {
+                               DRM_ERROR("Cannot set rate (%dHz) for gdp\n",
+                                               rate);
+                               return 1;
+                       }
+
+                       if (clk_prepare_enable(gdp->clk_pix)) {
+                               DRM_ERROR("Failed to prepare/enable gdp\n");
+                               return 1;
+                       }
+               }
+       }
+
+       return 0;
+}
+
+/**
+ * sti_gdp_commit_layer
+ * @lay: gdp layer
+ *
+ * Update the NVN field of the 'right' field of the current GDP node (being
+ * used by the HW) with the address of the updated ('free') top field GDP node.
+ * - In interlaced mode the 'right' field is the bottom field as we update
+ *   frames starting from their top field
+ * - In progressive mode, we update both bottom and top fields which are
+ *   equal nodes.
+ * At the next VSYNC, the updated node list will be used by the HW.
+ *
+ * RETURNS:
+ * 0 on success.
+ */
+static int sti_gdp_commit_layer(struct sti_layer *layer)
+{
+       struct sti_gdp_node_list *updated_list = sti_gdp_get_free_nodes(layer);
+       struct sti_gdp_node *updated_top_node = updated_list->top_field;
+       struct sti_gdp_node *updated_btm_node = updated_list->btm_field;
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       u32 dma_updated_top = updated_list->top_field_paddr;
+       u32 dma_updated_btm = updated_list->btm_field_paddr;
+       struct sti_gdp_node_list *curr_list = sti_gdp_get_current_nodes(layer);
+
+       dev_dbg(layer->dev, "%s %s top/btm_node:0x%p/0x%p\n", __func__,
+                       sti_layer_to_str(layer),
+                       updated_top_node, updated_btm_node);
+       dev_dbg(layer->dev, "Current NVN:0x%X\n",
+               readl(layer->regs + GAM_GDP_NVN_OFFSET));
+       dev_dbg(layer->dev, "Posted buff: %lx current buff: %x\n",
+               (unsigned long)layer->paddr,
+               readl(layer->regs + GAM_GDP_PML_OFFSET));
+
+       if (curr_list == NULL) {
+               /* First update or invalid node should directly write in the
+                * hw register */
+               DRM_DEBUG_DRIVER("%s first update (or invalid node)",
+                               sti_layer_to_str(layer));
+
+               writel(gdp->is_curr_top == true ?
+                               dma_updated_btm : dma_updated_top,
+                               layer->regs + GAM_GDP_NVN_OFFSET);
+               return 0;
+       }
+
+       if (layer->mode->flags & DRM_MODE_FLAG_INTERLACE) {
+               if (gdp->is_curr_top == true) {
+                       /* Do not update in the middle of the frame, but
+                        * postpone the update after the bottom field has
+                        * been displayed */
+                       curr_list->btm_field->gam_gdp_nvn = dma_updated_top;
+               } else {
+                       /* Direct update to avoid one frame delay */
+                       writel(dma_updated_top,
+                               layer->regs + GAM_GDP_NVN_OFFSET);
+               }
+       } else {
+               /* Direct update for progressive to avoid one frame delay */
+               writel(dma_updated_top, layer->regs + GAM_GDP_NVN_OFFSET);
+       }
+
+       return 0;
+}
+
+/**
+ * sti_gdp_disable_layer
+ * @lay: gdp layer
+ *
+ * Disable a GDP.
+ *
+ * RETURNS:
+ * 0 on success.
+ */
+static int sti_gdp_disable_layer(struct sti_layer *layer)
+{
+       unsigned int i;
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       struct sti_compositor *compo = dev_get_drvdata(layer->dev);
+
+       DRM_DEBUG_DRIVER("%s\n", sti_layer_to_str(layer));
+
+       /* Set the nodes as 'to be ignored on mixer' */
+       for (i = 0; i < GDP_NODE_NB_BANK; i++) {
+               gdp->node_list[i].top_field->gam_gdp_ppt |= GAM_GDP_PPT_IGNORE;
+               gdp->node_list[i].btm_field->gam_gdp_ppt |= GAM_GDP_PPT_IGNORE;
+       }
+
+       if (sti_vtg_unregister_client(layer->mixer_id == STI_MIXER_MAIN ?
+                       compo->vtg_main : compo->vtg_aux, &gdp->vtg_field_nb))
+               DRM_DEBUG_DRIVER("Warning: cannot unregister VTG notifier\n");
+
+       if (gdp->clk_pix)
+               clk_disable_unprepare(gdp->clk_pix);
+
+       return 0;
+}
+
+/**
+ * sti_gdp_field_cb
+ * @nb: notifier block
+ * @event: event message
+ * @data: private data
+ *
+ * Handle VTG top field and bottom field event.
+ *
+ * RETURNS:
+ * 0 on success.
+ */
+int sti_gdp_field_cb(struct notifier_block *nb,
+               unsigned long event, void *data)
+{
+       struct sti_gdp *gdp = container_of(nb, struct sti_gdp, vtg_field_nb);
+
+       switch (event) {
+       case VTG_TOP_FIELD_EVENT:
+               gdp->is_curr_top = true;
+               break;
+       case VTG_BOTTOM_FIELD_EVENT:
+               gdp->is_curr_top = false;
+               break;
+       default:
+               DRM_ERROR("unsupported event: %lu\n", event);
+               break;
+       }
+
+       return 0;
+}
+
+static void sti_gdp_init(struct sti_layer *layer)
+{
+       struct sti_gdp *gdp = to_sti_gdp(layer);
+       struct device_node *np = layer->dev->of_node;
+       dma_addr_t dma_addr;
+       void *base;
+       unsigned int i, size;
+
+       /* Allocate all the nodes within a single memory page */
+       size = sizeof(struct sti_gdp_node) *
+           GDP_NODE_PER_FIELD * GDP_NODE_NB_BANK;
+       base = dma_alloc_writecombine(layer->dev,
+                       size, &dma_addr, GFP_KERNEL | GFP_DMA);
+
+       if (!base) {
+               DRM_ERROR("Failed to allocate memory for GDP node\n");
+               return;
+       }
+       memset(base, 0, size);
+
+       for (i = 0; i < GDP_NODE_NB_BANK; i++) {
+               if (dma_addr & 0xF) {
+                       DRM_ERROR("Mem alignment failed\n");
+                       return;
+               }
+               gdp->node_list[i].top_field = base;
+               gdp->node_list[i].top_field_paddr = dma_addr;
+
+               DRM_DEBUG_DRIVER("node[%d].top_field=%p\n", i, base);
+               base += sizeof(struct sti_gdp_node);
+               dma_addr += sizeof(struct sti_gdp_node);
+
+               if (dma_addr & 0xF) {
+                       DRM_ERROR("Mem alignment failed\n");
+                       return;
+               }
+               gdp->node_list[i].btm_field = base;
+               gdp->node_list[i].btm_field_paddr = dma_addr;
+               DRM_DEBUG_DRIVER("node[%d].btm_field=%p\n", i, base);
+               base += sizeof(struct sti_gdp_node);
+               dma_addr += sizeof(struct sti_gdp_node);
+       }
+
+       if (of_device_is_compatible(np, "st,stih407-compositor")) {
+               /* GDP of STiH407 chip have its own pixel clock */
+               char *clk_name;
+
+               switch (layer->desc) {
+               case STI_GDP_0:
+                       clk_name = "pix_gdp1";
+                       break;
+               case STI_GDP_1:
+                       clk_name = "pix_gdp2";
+                       break;
+               case STI_GDP_2:
+                       clk_name = "pix_gdp3";
+                       break;
+               case STI_GDP_3:
+                       clk_name = "pix_gdp4";
+                       break;
+               default:
+                       DRM_ERROR("GDP id not recognized\n");
+                       return;
+               }
+
+               gdp->clk_pix = devm_clk_get(layer->dev, clk_name);
+               if (IS_ERR(gdp->clk_pix))
+                       DRM_ERROR("Cannot get %s clock\n", clk_name);
+
+               gdp->clk_main_parent = devm_clk_get(layer->dev, "main_parent");
+               if (IS_ERR(gdp->clk_main_parent))
+                       DRM_ERROR("Cannot get main_parent clock\n");
+
+               gdp->clk_aux_parent = devm_clk_get(layer->dev, "aux_parent");
+               if (IS_ERR(gdp->clk_aux_parent))
+                       DRM_ERROR("Cannot get aux_parent clock\n");
+       }
+}
+
+static const struct sti_layer_funcs gdp_ops = {
+       .get_formats = sti_gdp_get_formats,
+       .get_nb_formats = sti_gdp_get_nb_formats,
+       .init = sti_gdp_init,
+       .prepare = sti_gdp_prepare_layer,
+       .commit = sti_gdp_commit_layer,
+       .disable = sti_gdp_disable_layer,
+};
+
+struct sti_layer *sti_gdp_create(struct device *dev, int id)
+{
+       struct sti_gdp *gdp;
+
+       gdp = devm_kzalloc(dev, sizeof(*gdp), GFP_KERNEL);
+       if (!gdp) {
+               DRM_ERROR("Failed to allocate memory for GDP\n");
+               return NULL;
+       }
+
+       gdp->layer.ops = &gdp_ops;
+       gdp->vtg_field_nb.notifier_call = sti_gdp_field_cb;
+
+       return (struct sti_layer *)gdp;
+}