// SPDX-License-Identifier: GPL-2.0+ /* * NVIDIA Tegra Video decoder driver * * Copyright (C) 2016-2017 Dmitry Osipenko * */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "vde.h" #define CREATE_TRACE_POINTS #include "trace.h" void tegra_vde_writel(struct tegra_vde *vde, u32 value, void __iomem *base, u32 offset) { trace_vde_writel(vde, base, offset, value); writel_relaxed(value, base + offset); } u32 tegra_vde_readl(struct tegra_vde *vde, void __iomem *base, u32 offset) { u32 value = readl_relaxed(base + offset); trace_vde_readl(vde, base, offset, value); return value; } void tegra_vde_set_bits(struct tegra_vde *vde, u32 mask, void __iomem *base, u32 offset) { u32 value = tegra_vde_readl(vde, base, offset); tegra_vde_writel(vde, value | mask, base, offset); } int tegra_vde_alloc_bo(struct tegra_vde *vde, struct tegra_vde_bo **ret_bo, enum dma_data_direction dma_dir, size_t size) { struct device *dev = vde->dev; struct tegra_vde_bo *bo; int err; bo = kzalloc(sizeof(*bo), GFP_KERNEL); if (!bo) return -ENOMEM; bo->vde = vde; bo->size = size; bo->dma_dir = dma_dir; bo->dma_attrs = DMA_ATTR_WRITE_COMBINE | DMA_ATTR_NO_KERNEL_MAPPING; if (!vde->domain) bo->dma_attrs |= DMA_ATTR_FORCE_CONTIGUOUS; bo->dma_cookie = dma_alloc_attrs(dev, bo->size, &bo->dma_handle, GFP_KERNEL, bo->dma_attrs); if (!bo->dma_cookie) { dev_err(dev, "Failed to allocate DMA buffer of size: %zu\n", bo->size); err = -ENOMEM; goto free_bo; } err = dma_get_sgtable_attrs(dev, &bo->sgt, bo->dma_cookie, bo->dma_handle, bo->size, bo->dma_attrs); if (err) { dev_err(dev, "Failed to get DMA buffer SG table: %d\n", err); goto free_attrs; } err = dma_map_sgtable(dev, &bo->sgt, bo->dma_dir, bo->dma_attrs); if (err) { dev_err(dev, "Failed to map DMA buffer SG table: %d\n", err); goto free_table; } if (vde->domain) { err = tegra_vde_iommu_map(vde, &bo->sgt, &bo->iova, bo->size); if (err) { dev_err(dev, "Failed to map DMA buffer IOVA: %d\n", err); goto unmap_sgtable; } bo->dma_addr = iova_dma_addr(&vde->iova, bo->iova); } else { bo->dma_addr = sg_dma_address(bo->sgt.sgl); } *ret_bo = bo; return 0; unmap_sgtable: dma_unmap_sgtable(dev, &bo->sgt, bo->dma_dir, bo->dma_attrs); free_table: sg_free_table(&bo->sgt); free_attrs: dma_free_attrs(dev, bo->size, bo->dma_cookie, bo->dma_handle, bo->dma_attrs); free_bo: kfree(bo); return err; } void tegra_vde_free_bo(struct tegra_vde_bo *bo) { struct tegra_vde *vde = bo->vde; struct device *dev = vde->dev; if (vde->domain) tegra_vde_iommu_unmap(vde, bo->iova); dma_unmap_sgtable(dev, &bo->sgt, bo->dma_dir, bo->dma_attrs); sg_free_table(&bo->sgt); dma_free_attrs(dev, bo->size, bo->dma_cookie, bo->dma_handle, bo->dma_attrs); kfree(bo); } static irqreturn_t tegra_vde_isr(int irq, void *data) { struct tegra_vde *vde = data; if (completion_done(&vde->decode_completion)) return IRQ_NONE; tegra_vde_set_bits(vde, 0, vde->frameid, 0x208); complete(&vde->decode_completion); return IRQ_HANDLED; } static __maybe_unused int tegra_vde_runtime_suspend(struct device *dev) { struct tegra_vde *vde = dev_get_drvdata(dev); int err; if (!dev->pm_domain) { err = tegra_powergate_power_off(TEGRA_POWERGATE_VDEC); if (err) { dev_err(dev, "Failed to power down HW: %d\n", err); return err; } } clk_disable_unprepare(vde->clk); reset_control_release(vde->rst); reset_control_release(vde->rst_mc); return 0; } static __maybe_unused int tegra_vde_runtime_resume(struct device *dev) { struct tegra_vde *vde = dev_get_drvdata(dev); int err; err = reset_control_acquire(vde->rst_mc); if (err) { dev_err(dev, "Failed to acquire mc reset: %d\n", err); return err; } err = reset_control_acquire(vde->rst); if (err) { dev_err(dev, "Failed to acquire reset: %d\n", err); goto release_mc_reset; } if (!dev->pm_domain) { err = tegra_powergate_sequence_power_up(TEGRA_POWERGATE_VDEC, vde->clk, vde->rst); if (err) { dev_err(dev, "Failed to power up HW : %d\n", err); goto release_reset; } } else { /* * tegra_powergate_sequence_power_up() leaves clocks enabled, * while GENPD not. */ err = clk_prepare_enable(vde->clk); if (err) { dev_err(dev, "Failed to enable clock: %d\n", err); goto release_reset; } } return 0; release_reset: reset_control_release(vde->rst); release_mc_reset: reset_control_release(vde->rst_mc); return err; } static int tegra_vde_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; struct tegra_vde *vde; int irq, err; vde = devm_kzalloc(dev, sizeof(*vde), GFP_KERNEL); if (!vde) return -ENOMEM; platform_set_drvdata(pdev, vde); vde->soc = of_device_get_match_data(&pdev->dev); vde->dev = dev; vde->sxe = devm_platform_ioremap_resource_byname(pdev, "sxe"); if (IS_ERR(vde->sxe)) return PTR_ERR(vde->sxe); vde->bsev = devm_platform_ioremap_resource_byname(pdev, "bsev"); if (IS_ERR(vde->bsev)) return PTR_ERR(vde->bsev); vde->mbe = devm_platform_ioremap_resource_byname(pdev, "mbe"); if (IS_ERR(vde->mbe)) return PTR_ERR(vde->mbe); vde->ppe = devm_platform_ioremap_resource_byname(pdev, "ppe"); if (IS_ERR(vde->ppe)) return PTR_ERR(vde->ppe); vde->mce = devm_platform_ioremap_resource_byname(pdev, "mce"); if (IS_ERR(vde->mce)) return PTR_ERR(vde->mce); vde->tfe = devm_platform_ioremap_resource_byname(pdev, "tfe"); if (IS_ERR(vde->tfe)) return PTR_ERR(vde->tfe); vde->ppb = devm_platform_ioremap_resource_byname(pdev, "ppb"); if (IS_ERR(vde->ppb)) return PTR_ERR(vde->ppb); vde->vdma = devm_platform_ioremap_resource_byname(pdev, "vdma"); if (IS_ERR(vde->vdma)) return PTR_ERR(vde->vdma); vde->frameid = devm_platform_ioremap_resource_byname(pdev, "frameid"); if (IS_ERR(vde->frameid)) return PTR_ERR(vde->frameid); vde->clk = devm_clk_get(dev, NULL); if (IS_ERR(vde->clk)) { err = PTR_ERR(vde->clk); dev_err(dev, "Could not get VDE clk %d\n", err); return err; } vde->rst = devm_reset_control_get_exclusive_released(dev, NULL); if (IS_ERR(vde->rst)) { err = PTR_ERR(vde->rst); dev_err(dev, "Could not get VDE reset %d\n", err); return err; } vde->rst_mc = devm_reset_control_get_optional_exclusive_released(dev, "mc"); if (IS_ERR(vde->rst_mc)) { err = PTR_ERR(vde->rst_mc); dev_err(dev, "Could not get MC reset %d\n", err); return err; } irq = platform_get_irq_byname(pdev, "sync-token"); if (irq < 0) return irq; err = devm_request_irq(dev, irq, tegra_vde_isr, 0, dev_name(dev), vde); if (err) { dev_err(dev, "Could not request IRQ %d\n", err); return err; } err = devm_tegra_core_dev_init_opp_table_common(dev); if (err) { dev_err(dev, "Could initialize OPP table %d\n", err); return err; } vde->iram_pool = of_gen_pool_get(dev->of_node, "iram", 0); if (!vde->iram_pool) { dev_err(dev, "Could not get IRAM pool\n"); return -EPROBE_DEFER; } vde->iram = gen_pool_dma_alloc(vde->iram_pool, gen_pool_size(vde->iram_pool), &vde->iram_lists_addr); if (!vde->iram) { dev_err(dev, "Could not reserve IRAM\n"); return -ENOMEM; } INIT_LIST_HEAD(&vde->map_list); mutex_init(&vde->map_lock); mutex_init(&vde->lock); init_completion(&vde->decode_completion); err = tegra_vde_iommu_init(vde); if (err) { dev_err(dev, "Failed to initialize IOMMU: %d\n", err); goto err_gen_free; } pm_runtime_enable(dev); pm_runtime_use_autosuspend(dev); pm_runtime_set_autosuspend_delay(dev, 300); /* * VDE partition may be left ON after bootloader, hence let's * power-cycle it in order to put hardware into a predictable lower * power state. */ err = pm_runtime_resume_and_get(dev); if (err) goto err_pm_runtime; pm_runtime_put(dev); err = tegra_vde_alloc_bo(vde, &vde->secure_bo, DMA_FROM_DEVICE, 4096); if (err) { dev_err(dev, "Failed to allocate secure BO: %d\n", err); goto err_pm_runtime; } err = tegra_vde_v4l2_init(vde); if (err) { dev_err(dev, "Failed to initialize V4L2: %d\n", err); goto err_free_secure_bo; } return 0; err_free_secure_bo: tegra_vde_free_bo(vde->secure_bo); err_pm_runtime: pm_runtime_dont_use_autosuspend(dev); pm_runtime_disable(dev); tegra_vde_iommu_deinit(vde); err_gen_free: gen_pool_free(vde->iram_pool, (unsigned long)vde->iram, gen_pool_size(vde->iram_pool)); return err; } static void tegra_vde_remove(struct platform_device *pdev) { struct tegra_vde *vde = platform_get_drvdata(pdev); struct device *dev = &pdev->dev; tegra_vde_v4l2_deinit(vde); tegra_vde_free_bo(vde->secure_bo); /* * As it increments RPM usage_count even on errors, we don't need to * check the returned code here. */ pm_runtime_get_sync(dev); pm_runtime_dont_use_autosuspend(dev); pm_runtime_disable(dev); /* * Balance RPM state, the VDE power domain is left ON and hardware * is clock-gated. It's safe to reboot machine now. */ pm_runtime_put_noidle(dev); clk_disable_unprepare(vde->clk); tegra_vde_dmabuf_cache_unmap_all(vde); tegra_vde_iommu_deinit(vde); gen_pool_free(vde->iram_pool, (unsigned long)vde->iram, gen_pool_size(vde->iram_pool)); } static void tegra_vde_shutdown(struct platform_device *pdev) { /* * On some devices bootloader isn't ready to a power-gated VDE on * a warm-reboot, machine will hang in that case. */ pm_runtime_get_sync(&pdev->dev); } static __maybe_unused int tegra_vde_pm_suspend(struct device *dev) { struct tegra_vde *vde = dev_get_drvdata(dev); int err; mutex_lock(&vde->lock); err = pm_runtime_force_suspend(dev); if (err < 0) return err; return 0; } static __maybe_unused int tegra_vde_pm_resume(struct device *dev) { struct tegra_vde *vde = dev_get_drvdata(dev); int err; err = pm_runtime_force_resume(dev); if (err < 0) return err; mutex_unlock(&vde->lock); return 0; } static const struct dev_pm_ops tegra_vde_pm_ops = { SET_RUNTIME_PM_OPS(tegra_vde_runtime_suspend, tegra_vde_runtime_resume, NULL) SET_SYSTEM_SLEEP_PM_OPS(tegra_vde_pm_suspend, tegra_vde_pm_resume) }; static const u32 tegra124_decoded_fmts[] = { /* TBD: T124 supports only a non-standard Tegra tiled format */ }; static const struct tegra_coded_fmt_desc tegra124_coded_fmts[] = { { .fourcc = V4L2_PIX_FMT_H264_SLICE, .frmsize = { .min_width = 16, .max_width = 1920, .step_width = 16, .min_height = 16, .max_height = 2032, .step_height = 16, }, .num_decoded_fmts = ARRAY_SIZE(tegra124_decoded_fmts), .decoded_fmts = tegra124_decoded_fmts, .decode_run = tegra_vde_h264_decode_run, .decode_wait = tegra_vde_h264_decode_wait, }, }; static const u32 tegra20_decoded_fmts[] = { V4L2_PIX_FMT_YUV420M, V4L2_PIX_FMT_YVU420M, }; static const struct tegra_coded_fmt_desc tegra20_coded_fmts[] = { { .fourcc = V4L2_PIX_FMT_H264_SLICE, .frmsize = { .min_width = 16, .max_width = 1920, .step_width = 16, .min_height = 16, .max_height = 2032, .step_height = 16, }, .num_decoded_fmts = ARRAY_SIZE(tegra20_decoded_fmts), .decoded_fmts = tegra20_decoded_fmts, .decode_run = tegra_vde_h264_decode_run, .decode_wait = tegra_vde_h264_decode_wait, }, }; static const struct tegra_vde_soc tegra124_vde_soc = { .supports_ref_pic_marking = true, .coded_fmts = tegra124_coded_fmts, .num_coded_fmts = ARRAY_SIZE(tegra124_coded_fmts), }; static const struct tegra_vde_soc tegra114_vde_soc = { .supports_ref_pic_marking = true, .coded_fmts = tegra20_coded_fmts, .num_coded_fmts = ARRAY_SIZE(tegra20_coded_fmts), }; static const struct tegra_vde_soc tegra30_vde_soc = { .supports_ref_pic_marking = false, .coded_fmts = tegra20_coded_fmts, .num_coded_fmts = ARRAY_SIZE(tegra20_coded_fmts), }; static const struct tegra_vde_soc tegra20_vde_soc = { .supports_ref_pic_marking = false, .coded_fmts = tegra20_coded_fmts, .num_coded_fmts = ARRAY_SIZE(tegra20_coded_fmts), }; static const struct of_device_id tegra_vde_of_match[] = { { .compatible = "nvidia,tegra124-vde", .data = &tegra124_vde_soc }, { .compatible = "nvidia,tegra114-vde", .data = &tegra114_vde_soc }, { .compatible = "nvidia,tegra30-vde", .data = &tegra30_vde_soc }, { .compatible = "nvidia,tegra20-vde", .data = &tegra20_vde_soc }, { }, }; MODULE_DEVICE_TABLE(of, tegra_vde_of_match); static struct platform_driver tegra_vde_driver = { .probe = tegra_vde_probe, .remove = tegra_vde_remove, .shutdown = tegra_vde_shutdown, .driver = { .name = "tegra-vde", .of_match_table = tegra_vde_of_match, .pm = &tegra_vde_pm_ops, }, }; module_platform_driver(tegra_vde_driver); MODULE_DESCRIPTION("NVIDIA Tegra Video Decoder driver"); MODULE_AUTHOR("Dmitry Osipenko "); MODULE_LICENSE("GPL");