These changes are the raw update to linux-4.4.6-rt14. Kernel sources
[kvmfornfv.git] / kernel / drivers / gpu / drm / nouveau / nvkm / engine / gr / gk20a.c
index 40ff5eb..b8758d3 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2014, NVIDIA CORPORATION. All rights reserved.
+ * Copyright (c) 2014-2015, NVIDIA CORPORATION. All rights reserved.
  *
  * Permission is hereby granted, free of charge, to any person obtaining a
  * copy of this software and associated documentation files (the "Software"),
 #include "gf100.h"
 #include "ctxgf100.h"
 
+#include <subdev/timer.h>
+
 #include <nvif/class.h>
 
-static struct nvkm_oclass
-gk20a_gr_sclass[] = {
-       { FERMI_TWOD_A, &nvkm_object_ofuncs },
-       { KEPLER_INLINE_TO_MEMORY_A, &nvkm_object_ofuncs },
-       { KEPLER_C, &gf100_fermi_ofuncs, gf100_gr_9097_omthds },
-       { KEPLER_COMPUTE_A, &nvkm_object_ofuncs, gf100_gr_90c0_omthds },
-       {}
+static void
+gk20a_gr_init_dtor(struct gf100_gr_pack *pack)
+{
+       vfree(pack);
+}
+
+struct gk20a_fw_av
+{
+       u32 addr;
+       u32 data;
+};
+
+static struct gf100_gr_pack *
+gk20a_gr_av_to_init(struct gf100_gr_fuc *fuc)
+{
+       struct gf100_gr_init *init;
+       struct gf100_gr_pack *pack;
+       const int nent = (fuc->size / sizeof(struct gk20a_fw_av));
+       int i;
+
+       pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1)));
+       if (!pack)
+               return ERR_PTR(-ENOMEM);
+
+       init = (void *)(pack + 2);
+
+       pack[0].init = init;
+
+       for (i = 0; i < nent; i++) {
+               struct gf100_gr_init *ent = &init[i];
+               struct gk20a_fw_av *av = &((struct gk20a_fw_av *)fuc->data)[i];
+
+               ent->addr = av->addr;
+               ent->data = av->data;
+               ent->count = 1;
+               ent->pitch = 1;
+       }
+
+       return pack;
+}
+
+struct gk20a_fw_aiv
+{
+       u32 addr;
+       u32 index;
+       u32 data;
 };
 
-struct nvkm_oclass *
-gk20a_gr_oclass = &(struct gf100_gr_oclass) {
-       .base.handle = NV_ENGINE(GR, 0xea),
-       .base.ofuncs = &(struct nvkm_ofuncs) {
-               .ctor = gf100_gr_ctor,
-               .dtor = gf100_gr_dtor,
-               .init = gk104_gr_init,
-               .fini = _nvkm_gr_fini,
-       },
-       .cclass = &gk20a_grctx_oclass,
-       .sclass = gk20a_gr_sclass,
-       .mmio = gk104_gr_pack_mmio,
+static struct gf100_gr_pack *
+gk20a_gr_aiv_to_init(struct gf100_gr_fuc *fuc)
+{
+       struct gf100_gr_init *init;
+       struct gf100_gr_pack *pack;
+       const int nent = (fuc->size / sizeof(struct gk20a_fw_aiv));
+       int i;
+
+       pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1)));
+       if (!pack)
+               return ERR_PTR(-ENOMEM);
+
+       init = (void *)(pack + 2);
+
+       pack[0].init = init;
+
+       for (i = 0; i < nent; i++) {
+               struct gf100_gr_init *ent = &init[i];
+               struct gk20a_fw_aiv *av = &((struct gk20a_fw_aiv *)fuc->data)[i];
+
+               ent->addr = av->addr;
+               ent->data = av->data;
+               ent->count = 1;
+               ent->pitch = 1;
+       }
+
+       return pack;
+}
+
+static struct gf100_gr_pack *
+gk20a_gr_av_to_method(struct gf100_gr_fuc *fuc)
+{
+       struct gf100_gr_init *init;
+       struct gf100_gr_pack *pack;
+       /* We don't suppose we will initialize more than 16 classes here... */
+       static const unsigned int max_classes = 16;
+       const int nent = (fuc->size / sizeof(struct gk20a_fw_av));
+       int i, classidx = 0;
+       u32 prevclass = 0;
+
+       pack = vzalloc((sizeof(*pack) * max_classes) +
+                      (sizeof(*init) * (nent + 1)));
+       if (!pack)
+               return ERR_PTR(-ENOMEM);
+
+       init = (void *)(pack + max_classes);
+
+       for (i = 0; i < nent; i++) {
+               struct gf100_gr_init *ent = &init[i];
+               struct gk20a_fw_av *av = &((struct gk20a_fw_av *)fuc->data)[i];
+               u32 class = av->addr & 0xffff;
+               u32 addr = (av->addr & 0xffff0000) >> 14;
+
+               if (prevclass != class) {
+                       pack[classidx].init = ent;
+                       pack[classidx].type = class;
+                       prevclass = class;
+                       if (++classidx >= max_classes) {
+                               vfree(pack);
+                               return ERR_PTR(-ENOSPC);
+                       }
+               }
+
+               ent->addr = addr;
+               ent->data = av->data;
+               ent->count = 1;
+               ent->pitch = 1;
+       }
+
+       return pack;
+}
+
+static int
+gk20a_gr_wait_mem_scrubbing(struct gf100_gr *gr)
+{
+       struct nvkm_subdev *subdev = &gr->base.engine.subdev;
+       struct nvkm_device *device = subdev->device;
+
+       if (nvkm_msec(device, 2000,
+               if (!(nvkm_rd32(device, 0x40910c) & 0x00000006))
+                       break;
+       ) < 0) {
+               nvkm_error(subdev, "FECS mem scrubbing timeout\n");
+               return -ETIMEDOUT;
+       }
+
+       if (nvkm_msec(device, 2000,
+               if (!(nvkm_rd32(device, 0x41a10c) & 0x00000006))
+                       break;
+       ) < 0) {
+               nvkm_error(subdev, "GPCCS mem scrubbing timeout\n");
+               return -ETIMEDOUT;
+       }
+
+       return 0;
+}
+
+static void
+gk20a_gr_set_hww_esr_report_mask(struct gf100_gr *gr)
+{
+       struct nvkm_device *device = gr->base.engine.subdev.device;
+       nvkm_wr32(device, 0x419e44, 0x1ffffe);
+       nvkm_wr32(device, 0x419e4c, 0x7f);
+}
+
+int
+gk20a_gr_init(struct gf100_gr *gr)
+{
+       struct nvkm_device *device = gr->base.engine.subdev.device;
+       const u32 magicgpc918 = DIV_ROUND_UP(0x00800000, gr->tpc_total);
+       u32 data[TPC_MAX / 8] = {};
+       u8  tpcnr[GPC_MAX];
+       int gpc, tpc;
+       int ret, i;
+
+       /* Clear SCC RAM */
+       nvkm_wr32(device, 0x40802c, 0x1);
+
+       gf100_gr_mmio(gr, gr->fuc_sw_nonctx);
+
+       ret = gk20a_gr_wait_mem_scrubbing(gr);
+       if (ret)
+               return ret;
+
+       ret = gf100_gr_wait_idle(gr);
+       if (ret)
+               return ret;
+
+       /* MMU debug buffer */
+       nvkm_wr32(device, 0x100cc8, nvkm_memory_addr(gr->unk4188b4) >> 8);
+       nvkm_wr32(device, 0x100ccc, nvkm_memory_addr(gr->unk4188b8) >> 8);
+
+       if (gr->func->init_gpc_mmu)
+               gr->func->init_gpc_mmu(gr);
+
+       /* Set the PE as stream master */
+       nvkm_mask(device, 0x503018, 0x1, 0x1);
+
+       /* Zcull init */
+       memset(data, 0x00, sizeof(data));
+       memcpy(tpcnr, gr->tpc_nr, sizeof(gr->tpc_nr));
+       for (i = 0, gpc = -1; i < gr->tpc_total; i++) {
+               do {
+                       gpc = (gpc + 1) % gr->gpc_nr;
+               } while (!tpcnr[gpc]);
+               tpc = gr->tpc_nr[gpc] - tpcnr[gpc]--;
+
+               data[i / 8] |= tpc << ((i % 8) * 4);
+       }
+
+       nvkm_wr32(device, GPC_BCAST(0x0980), data[0]);
+       nvkm_wr32(device, GPC_BCAST(0x0984), data[1]);
+       nvkm_wr32(device, GPC_BCAST(0x0988), data[2]);
+       nvkm_wr32(device, GPC_BCAST(0x098c), data[3]);
+
+       for (gpc = 0; gpc < gr->gpc_nr; gpc++) {
+               nvkm_wr32(device, GPC_UNIT(gpc, 0x0914),
+                         gr->magic_not_rop_nr << 8 | gr->tpc_nr[gpc]);
+               nvkm_wr32(device, GPC_UNIT(gpc, 0x0910), 0x00040000 |
+                         gr->tpc_total);
+               nvkm_wr32(device, GPC_UNIT(gpc, 0x0918), magicgpc918);
+       }
+
+       nvkm_wr32(device, GPC_BCAST(0x3fd4), magicgpc918);
+
+       /* Enable FIFO access */
+       nvkm_wr32(device, 0x400500, 0x00010001);
+
+       /* Enable interrupts */
+       nvkm_wr32(device, 0x400100, 0xffffffff);
+       nvkm_wr32(device, 0x40013c, 0xffffffff);
+
+       /* Enable FECS error interrupts */
+       nvkm_wr32(device, 0x409c24, 0x000f0000);
+
+       /* Enable hardware warning exceptions */
+       nvkm_wr32(device, 0x404000, 0xc0000000);
+       nvkm_wr32(device, 0x404600, 0xc0000000);
+
+       if (gr->func->set_hww_esr_report_mask)
+               gr->func->set_hww_esr_report_mask(gr);
+
+       /* Enable TPC exceptions per GPC */
+       nvkm_wr32(device, 0x419d0c, 0x2);
+       nvkm_wr32(device, 0x41ac94, (((1 << gr->tpc_total) - 1) & 0xff) << 16);
+
+       /* Reset and enable all exceptions */
+       nvkm_wr32(device, 0x400108, 0xffffffff);
+       nvkm_wr32(device, 0x400138, 0xffffffff);
+       nvkm_wr32(device, 0x400118, 0xffffffff);
+       nvkm_wr32(device, 0x400130, 0xffffffff);
+       nvkm_wr32(device, 0x40011c, 0xffffffff);
+       nvkm_wr32(device, 0x400134, 0xffffffff);
+
+       gf100_gr_zbc_init(gr);
+
+       return gf100_gr_init_ctxctl(gr);
+}
+
+void
+gk20a_gr_dtor(struct gf100_gr *gr)
+{
+       gk20a_gr_init_dtor(gr->fuc_method);
+       gk20a_gr_init_dtor(gr->fuc_bundle);
+       gk20a_gr_init_dtor(gr->fuc_sw_ctx);
+       gk20a_gr_init_dtor(gr->fuc_sw_nonctx);
+}
+
+int
+gk20a_gr_new_(const struct gf100_gr_func *func, struct nvkm_device *device,
+             int index, struct nvkm_gr **pgr)
+{
+       struct gf100_gr_fuc fuc;
+       struct gf100_gr *gr;
+       int ret;
+
+       if (!(gr = kzalloc(sizeof(*gr), GFP_KERNEL)))
+               return -ENOMEM;
+       *pgr = &gr->base;
+
+       ret = gf100_gr_ctor(func, device, index, gr);
+       if (ret)
+               return ret;
+
+       ret = gf100_gr_ctor_fw(gr, "sw_nonctx", &fuc);
+       if (ret)
+               return ret;
+       gr->fuc_sw_nonctx = gk20a_gr_av_to_init(&fuc);
+       gf100_gr_dtor_fw(&fuc);
+       if (IS_ERR(gr->fuc_sw_nonctx))
+               return PTR_ERR(gr->fuc_sw_nonctx);
+
+       ret = gf100_gr_ctor_fw(gr, "sw_ctx", &fuc);
+       if (ret)
+               return ret;
+       gr->fuc_sw_ctx = gk20a_gr_aiv_to_init(&fuc);
+       gf100_gr_dtor_fw(&fuc);
+       if (IS_ERR(gr->fuc_sw_ctx))
+               return PTR_ERR(gr->fuc_sw_ctx);
+
+       ret = gf100_gr_ctor_fw(gr, "sw_bundle_init", &fuc);
+       if (ret)
+               return ret;
+       gr->fuc_bundle = gk20a_gr_av_to_init(&fuc);
+       gf100_gr_dtor_fw(&fuc);
+       if (IS_ERR(gr->fuc_bundle))
+               return PTR_ERR(gr->fuc_bundle);
+
+       ret = gf100_gr_ctor_fw(gr, "sw_method_init", &fuc);
+       if (ret)
+               return ret;
+       gr->fuc_method = gk20a_gr_av_to_method(&fuc);
+       gf100_gr_dtor_fw(&fuc);
+       if (IS_ERR(gr->fuc_method))
+               return PTR_ERR(gr->fuc_method);
+
+       return 0;
+}
+
+static const struct gf100_gr_func
+gk20a_gr = {
+       .dtor = gk20a_gr_dtor,
+       .init = gk20a_gr_init,
+       .set_hww_esr_report_mask = gk20a_gr_set_hww_esr_report_mask,
        .ppc_nr = 1,
-}.base;
+       .grctx = &gk20a_grctx,
+       .sclass = {
+               { -1, -1, FERMI_TWOD_A },
+               { -1, -1, KEPLER_INLINE_TO_MEMORY_A },
+               { -1, -1, KEPLER_C, &gf100_fermi },
+               { -1, -1, KEPLER_COMPUTE_A },
+               {}
+       }
+};
+
+int
+gk20a_gr_new(struct nvkm_device *device, int index, struct nvkm_gr **pgr)
+{
+       return gk20a_gr_new_(&gk20a_gr, device, index, pgr);
+}