// SPDX-License-Identifier: GPL-2.0 /* Copyright (c) 2012,2017-2020 The Linux Foundation. All rights reserved. * * Description: CoreSight Trace Memory Controller driver */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "coresight-priv.h" #include "coresight-tmc.h" #include "coresight-common.h" DEFINE_CORESIGHT_DEVLIST(etb_devs, "tmc_etb"); DEFINE_CORESIGHT_DEVLIST(etf_devs, "tmc_etf"); DEFINE_CORESIGHT_DEVLIST(etr_devs, "tmc_etr"); void tmc_wait_for_tmcready(struct tmc_drvdata *drvdata) { /* Ensure formatter, unformatter and hardware fifo are empty */ if (coresight_timeout(drvdata->base, TMC_STS, TMC_STS_TMCREADY_BIT, 1)) { dev_err(&drvdata->csdev->dev, "timeout while waiting for TMC to be Ready\n"); } } void tmc_flush_and_stop(struct tmc_drvdata *drvdata) { u32 ffcr; ffcr = readl_relaxed(drvdata->base + TMC_FFCR); ffcr |= TMC_FFCR_STOP_ON_FLUSH; writel_relaxed(ffcr, drvdata->base + TMC_FFCR); ffcr |= BIT(TMC_FFCR_FLUSHMAN_BIT); writel_relaxed(ffcr, drvdata->base + TMC_FFCR); /* Ensure flush completes */ if (coresight_timeout(drvdata->base, TMC_FFCR, TMC_FFCR_FLUSHMAN_BIT, 0)) { dev_err(&drvdata->csdev->dev, "timeout while waiting for completion of Manual Flush\n"); } tmc_wait_for_tmcready(drvdata); } void tmc_enable_hw(struct tmc_drvdata *drvdata) { drvdata->enable = true; writel_relaxed(TMC_CTL_CAPT_EN, drvdata->base + TMC_CTL); } EXPORT_SYMBOL(tmc_enable_hw); void tmc_disable_hw(struct tmc_drvdata *drvdata) { drvdata->enable = false; writel_relaxed(0x0, drvdata->base + TMC_CTL); } u32 tmc_get_memwidth_mask(struct tmc_drvdata *drvdata) { u32 mask = 0; /* * When moving RRP or an offset address forward, the new values must * be byte-address aligned to the width of the trace memory databus * _and_ to a frame boundary (16 byte), whichever is the biggest. For * example, for 32-bit, 64-bit and 128-bit wide trace memory, the four * LSBs must be 0s. For 256-bit wide trace memory, the five LSBs must * be 0s. */ switch (drvdata->memwidth) { case TMC_MEM_INTF_WIDTH_32BITS: /* fallthrough */ case TMC_MEM_INTF_WIDTH_64BITS: /* fallthrough */ case TMC_MEM_INTF_WIDTH_128BITS: mask = GENMASK(31, 4); break; case TMC_MEM_INTF_WIDTH_256BITS: mask = GENMASK(31, 5); break; } return mask; } static int tmc_read_prepare(struct tmc_drvdata *drvdata) { int ret = 0; if (!drvdata->enable) return -EPERM; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: ret = tmc_read_prepare_etb(drvdata); break; case TMC_CONFIG_TYPE_ETR: ret = tmc_read_prepare_etr(drvdata); break; default: ret = -EINVAL; } if (!ret) dev_dbg(&drvdata->csdev->dev, "TMC read start\n"); return ret; } static int tmc_read_unprepare(struct tmc_drvdata *drvdata) { int ret = 0; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: ret = tmc_read_unprepare_etb(drvdata); break; case TMC_CONFIG_TYPE_ETR: ret = tmc_read_unprepare_etr(drvdata); break; default: ret = -EINVAL; } if (!ret) dev_dbg(&drvdata->csdev->dev, "TMC read end\n"); return ret; } static int tmc_open(struct inode *inode, struct file *file) { int ret; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); ret = tmc_read_prepare(drvdata); if (ret) return ret; nonseekable_open(inode, file); dev_dbg(&drvdata->csdev->dev, "%s: successfully opened\n", __func__); return 0; } static inline ssize_t tmc_get_sysfs_trace(struct tmc_drvdata *drvdata, loff_t pos, size_t len, char **bufpp) { switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: return tmc_etb_get_sysfs_trace(drvdata, pos, len, bufpp); case TMC_CONFIG_TYPE_ETR: return tmc_etr_get_sysfs_trace(drvdata, pos, len, bufpp); } return -EINVAL; } static ssize_t tmc_read(struct file *file, char __user *data, size_t len, loff_t *ppos) { char *bufp; ssize_t actual; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); mutex_lock(&drvdata->mem_lock); actual = tmc_get_sysfs_trace(drvdata, *ppos, len, &bufp); if (actual <= 0) { mutex_unlock(&drvdata->mem_lock); return 0; } if (copy_to_user(data, bufp, actual)) { dev_dbg(&drvdata->csdev->dev, "%s: copy_to_user failed\n", __func__); mutex_unlock(&drvdata->mem_lock); return -EFAULT; } *ppos += actual; dev_dbg(&drvdata->csdev->dev, "%zu bytes copied\n", actual); mutex_unlock(&drvdata->mem_lock); return actual; } static int tmc_release(struct inode *inode, struct file *file) { int ret; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); ret = tmc_read_unprepare(drvdata); if (ret) return ret; dev_dbg(&drvdata->csdev->dev, "%s: released\n", __func__); return 0; } static const struct file_operations tmc_fops = { .owner = THIS_MODULE, .open = tmc_open, .read = tmc_read, .release = tmc_release, .llseek = no_llseek, }; static enum tmc_mem_intf_width tmc_get_memwidth(u32 devid) { enum tmc_mem_intf_width memwidth; /* * Excerpt from the TRM: * * DEVID::MEMWIDTH[10:8] * 0x2 Memory interface databus is 32 bits wide. * 0x3 Memory interface databus is 64 bits wide. * 0x4 Memory interface databus is 128 bits wide. * 0x5 Memory interface databus is 256 bits wide. */ switch (BMVAL(devid, 8, 10)) { case 0x2: memwidth = TMC_MEM_INTF_WIDTH_32BITS; break; case 0x3: memwidth = TMC_MEM_INTF_WIDTH_64BITS; break; case 0x4: memwidth = TMC_MEM_INTF_WIDTH_128BITS; break; case 0x5: memwidth = TMC_MEM_INTF_WIDTH_256BITS; break; default: memwidth = 0; } return memwidth; } #define coresight_tmc_reg(name, offset) \ coresight_simple_reg32(struct tmc_drvdata, name, offset) #define coresight_tmc_reg64(name, lo_off, hi_off) \ coresight_simple_reg64(struct tmc_drvdata, name, lo_off, hi_off) coresight_tmc_reg(rsz, TMC_RSZ); coresight_tmc_reg(sts, TMC_STS); coresight_tmc_reg(trg, TMC_TRG); coresight_tmc_reg(ctl, TMC_CTL); coresight_tmc_reg(ffsr, TMC_FFSR); coresight_tmc_reg(ffcr, TMC_FFCR); coresight_tmc_reg(mode, TMC_MODE); coresight_tmc_reg(pscr, TMC_PSCR); coresight_tmc_reg(axictl, TMC_AXICTL); coresight_tmc_reg(authstatus, TMC_AUTHSTATUS); coresight_tmc_reg(devid, CORESIGHT_DEVID); coresight_tmc_reg64(rrp, TMC_RRP, TMC_RRPHI); coresight_tmc_reg64(rwp, TMC_RWP, TMC_RWPHI); coresight_tmc_reg64(dba, TMC_DBALO, TMC_DBAHI); static struct attribute *coresight_tmc_mgmt_attrs[] = { &dev_attr_rsz.attr, &dev_attr_sts.attr, &dev_attr_rrp.attr, &dev_attr_rwp.attr, &dev_attr_trg.attr, &dev_attr_ctl.attr, &dev_attr_ffsr.attr, &dev_attr_ffcr.attr, &dev_attr_mode.attr, &dev_attr_pscr.attr, &dev_attr_devid.attr, &dev_attr_dba.attr, &dev_attr_axictl.attr, &dev_attr_authstatus.attr, NULL, }; static ssize_t trigger_cntr_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); unsigned long val = drvdata->trigger_cntr; return sprintf(buf, "%#lx\n", val); } static ssize_t trigger_cntr_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { int ret; unsigned long val; struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); ret = kstrtoul(buf, 16, &val); if (ret) return ret; drvdata->trigger_cntr = val; return size; } static DEVICE_ATTR_RW(trigger_cntr); static ssize_t buffer_size_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); return scnprintf(buf, PAGE_SIZE, "%#x\n", drvdata->size); } static ssize_t buffer_size_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { int ret; unsigned long val; struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); if (drvdata->enable) { pr_err("ETR is in use, disable it to change the mem_size\n"); return -EINVAL; } /* Only permitted for TMC-ETRs */ if (drvdata->config_type != TMC_CONFIG_TYPE_ETR) return -EPERM; ret = kstrtoul(buf, 0, &val); if (ret) return ret; /* The buffer size should be page aligned */ if (val & (PAGE_SIZE - 1)) return -EINVAL; drvdata->size = val; return size; } static DEVICE_ATTR_RW(buffer_size); static ssize_t block_size_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); uint32_t val = 0; if (drvdata->byte_cntr) val = drvdata->byte_cntr->block_size; return scnprintf(buf, PAGE_SIZE, "%d\n", val); } static ssize_t block_size_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); unsigned long val; if (kstrtoul(buf, 0, &val)) return -EINVAL; if (!drvdata->byte_cntr) return -EINVAL; if (val && val < 4096) { pr_err("Assign minimum block size of 4096 bytes\n"); return -EINVAL; } mutex_lock(&drvdata->byte_cntr->byte_cntr_lock); drvdata->byte_cntr->block_size = val; mutex_unlock(&drvdata->byte_cntr->byte_cntr_lock); return size; } static DEVICE_ATTR_RW(block_size); static ssize_t out_mode_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); return scnprintf(buf, PAGE_SIZE, "%s\n", str_tmc_etr_out_mode[drvdata->out_mode]); } static ssize_t out_mode_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); char str[16] = ""; int ret; if (strlen(buf) >= 16) return -EINVAL; if (sscanf(buf, "%16s", str) != 1) return -EINVAL; ret = tmc_etr_switch_mode(drvdata, str); return ret ? ret : size; } static DEVICE_ATTR_RW(out_mode); static ssize_t available_out_modes_show(struct device *dev, struct device_attribute *attr, char *buf) { ssize_t len = 0; int i; for (i = 0; i < ARRAY_SIZE(str_tmc_etr_out_mode); i++) len += scnprintf(buf + len, PAGE_SIZE - len, "%s ", str_tmc_etr_out_mode[i]); len += scnprintf(buf + len, PAGE_SIZE - len, "\n"); return len; } static DEVICE_ATTR_RO(available_out_modes); static struct attribute *coresight_tmc_etf_attrs[] = { &dev_attr_trigger_cntr.attr, NULL, }; static struct attribute *coresight_tmc_etr_attrs[] = { &dev_attr_trigger_cntr.attr, &dev_attr_buffer_size.attr, &dev_attr_block_size.attr, &dev_attr_out_mode.attr, &dev_attr_available_out_modes.attr, NULL, }; static const struct attribute_group coresight_tmc_etf_group = { .attrs = coresight_tmc_etf_attrs, }; static const struct attribute_group coresight_tmc_etr_group = { .attrs = coresight_tmc_etr_attrs, }; static const struct attribute_group coresight_tmc_mgmt_group = { .attrs = coresight_tmc_mgmt_attrs, .name = "mgmt", }; static const struct attribute_group *coresight_tmc_etf_groups[] = { &coresight_tmc_etf_group, &coresight_tmc_mgmt_group, NULL, }; static const struct attribute_group *coresight_tmc_etr_groups[] = { &coresight_tmc_etr_group, &coresight_tmc_mgmt_group, NULL, }; static inline bool tmc_etr_can_use_sg(struct device *dev) { return fwnode_property_present(dev->fwnode, "arm,scatter-gather"); } static inline bool tmc_etr_has_non_secure_access(struct tmc_drvdata *drvdata) { u32 auth = readl_relaxed(drvdata->base + TMC_AUTHSTATUS); return (auth & TMC_AUTH_NSID_MASK) == 0x3; } /* Detect and initialise the capabilities of a TMC ETR */ static int tmc_etr_setup_caps(struct device *parent, u32 devid, void *dev_caps) { int rc; u32 dma_mask = 0; struct tmc_drvdata *drvdata = dev_get_drvdata(parent); if (!tmc_etr_has_non_secure_access(drvdata)) return -EACCES; /* Set the unadvertised capabilities */ tmc_etr_init_caps(drvdata, (u32)(unsigned long)dev_caps); if (!(devid & TMC_DEVID_NOSCAT) && tmc_etr_can_use_sg(parent)) tmc_etr_set_cap(drvdata, TMC_ETR_SG); /* Check if the AXI address width is available */ if (devid & TMC_DEVID_AXIAW_VALID) dma_mask = ((devid >> TMC_DEVID_AXIAW_SHIFT) & TMC_DEVID_AXIAW_MASK); /* * Unless specified in the device configuration, ETR uses a 40-bit * AXI master in place of the embedded SRAM of ETB/ETF. */ switch (dma_mask) { case 32: case 40: case 44: case 48: case 52: dev_info(parent, "Detected dma mask %dbits\n", dma_mask); break; default: dma_mask = 40; } rc = dma_set_mask_and_coherent(parent, DMA_BIT_MASK(dma_mask)); if (rc) dev_err(parent, "Failed to setup DMA mask: %d\n", rc); return rc; } static u32 tmc_etr_get_default_buffer_size(struct device *dev) { u32 size; if (fwnode_property_read_u32(dev->fwnode, "arm,buffer-size", &size)) size = SZ_1M; return size; } static void tmc_get_q6_etr_region(struct device *dev) { struct device_node *np; struct reserved_mem *rmem; struct tmc_drvdata *drvdata = dev_get_drvdata(dev); np = of_parse_phandle(dev->of_node, "memory-region", 0); if (!np) { dev_info(dev, "No Q6 ETR memory region specified\n"); return; } rmem = of_reserved_mem_lookup(np); of_node_put(np); if (!rmem) { dev_err(dev, "unable to acquire Q6 ETR memory-region\n"); return; } drvdata->q6_etr_vaddr = devm_ioremap_nocache(dev, rmem->base, rmem->size); if (drvdata->q6_etr_vaddr) { drvdata->q6_etr_paddr = rmem->base; drvdata->q6_size = rmem->size; } } static irqreturn_t etr_handler(int irq, void *data) { struct tmc_drvdata *drvdata = data; unsigned long flags; spin_lock_irqsave(&drvdata->spinlock, flags); if (drvdata->out_mode == TMC_ETR_OUT_MODE_Q6MEM_STREAM) { CS_UNLOCK(drvdata->base); atomic_add(PAGES_PER_DATA, &drvdata->seq_no); if (atomic_read(&drvdata->seq_no) > COMP_PAGES_PER_DATA) tmc_disable_hw(drvdata); else tmc_enable_hw(drvdata); schedule_work(&drvdata->qld_stream_work); CS_LOCK(drvdata->base); } spin_unlock_irqrestore(&drvdata->spinlock, flags); return IRQ_HANDLED; } struct tmc_drvdata *tmc_drvdata_stream; EXPORT_SYMBOL(tmc_drvdata_stream); static int tmc_probe(struct amba_device *adev, const struct amba_id *id) { int ret = 0; u32 devid; void __iomem *base; struct device *dev = &adev->dev; struct coresight_platform_data *pdata = NULL; struct tmc_drvdata *drvdata; struct resource *res = &adev->res; struct coresight_desc desc = { 0 }; struct coresight_dev_list *dev_list = NULL; struct coresight_cti_data *ctidata; int irq; u32 in_funnel_addr[2]; ret = -ENOMEM; drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL); if (!drvdata) goto out; drvdata->etr_usb_clk = devm_clk_get(&adev->dev, "etr_usb_clk"); if (!IS_ERR(drvdata->etr_usb_clk)) { ret = clk_prepare_enable(drvdata->etr_usb_clk); if (ret) return ret; } dev_set_drvdata(dev, drvdata); /* Validity for the resource is already checked by the AMBA core */ base = devm_ioremap_resource(dev, res); if (IS_ERR(base)) { ret = PTR_ERR(base); goto out; } drvdata->base = base; spin_lock_init(&drvdata->spinlock); mutex_init(&drvdata->mem_lock); devid = readl_relaxed(drvdata->base + CORESIGHT_DEVID); drvdata->config_type = BMVAL(devid, 6, 7); drvdata->memwidth = tmc_get_memwidth(devid); /* This device is not associated with a session */ drvdata->pid = -1; if (drvdata->config_type == TMC_CONFIG_TYPE_ETR) { drvdata->out_mode = TMC_ETR_OUT_MODE_MEM; drvdata->size = tmc_etr_get_default_buffer_size(dev); tmc_get_q6_etr_region(dev); } else { drvdata->size = readl_relaxed(drvdata->base + TMC_RSZ) * 4; } ret = of_get_coresight_csr_name(adev->dev.of_node, &drvdata->csr_name); if (ret) dev_dbg(dev, "No csr data\n"); else { drvdata->csr = coresight_csr_get(drvdata->csr_name); if (IS_ERR(drvdata->csr)) { dev_dbg(dev, "failed to get csr, defer probe\n"); return -EPROBE_DEFER; } } ctidata = of_get_coresight_cti_data(dev, adev->dev.of_node); if (IS_ERR(ctidata)) { dev_err(dev, "invalid cti data\n"); } else if (ctidata && ctidata->nr_ctis == 2) { drvdata->cti_flush = coresight_cti_get(ctidata->names[0]); if (IS_ERR(drvdata->cti_flush)) { dev_err(dev, "failed to get flush cti, defer probe\n"); return -EPROBE_DEFER; } drvdata->cti_reset = coresight_cti_get(ctidata->names[1]); if (IS_ERR(drvdata->cti_reset)) { dev_err(dev, "failed to get reset cti, defer probe\n"); return -EPROBE_DEFER; } } if (!of_property_read_u32_array(adev->dev.of_node, "funnel-address", in_funnel_addr, ARRAY_SIZE(in_funnel_addr))) { drvdata->in_funnel_base = ioremap(in_funnel_addr[0], in_funnel_addr[1]); if(IS_ERR(drvdata->in_funnel_base)) dev_err(dev, "unable to ioremap the funnel address\n"); } desc.dev = dev; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: desc.type = CORESIGHT_DEV_TYPE_SINK; desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; desc.groups = coresight_tmc_etf_groups; desc.ops = &tmc_etb_cs_ops; dev_list = &etb_devs; break; case TMC_CONFIG_TYPE_ETR: desc.type = CORESIGHT_DEV_TYPE_SINK; desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; desc.groups = coresight_tmc_etr_groups; desc.ops = &tmc_etr_cs_ops; ret = tmc_etr_setup_caps(dev, devid, coresight_get_uci_data(id)); if (ret) goto out; drvdata->byte_cntr = NULL; irq = of_irq_get_byname(adev->dev.of_node, "byte-cntr-irq"); if (irq > 0) { if (devm_request_irq(dev, irq, etr_handler, IRQF_TRIGGER_RISING | IRQF_SHARED, "tmc-etr", drvdata)) dev_err(dev, "Byte_cntr interrupt registration failed\n"); } atomic_set(&drvdata->completed_seq_no, 0); atomic_set(&drvdata->seq_no, 0); tmc_drvdata_stream = drvdata; ret = tmc_etr_bam_init(adev, drvdata); if (ret) goto out; idr_init(&drvdata->idr); mutex_init(&drvdata->idr_mutex); dev_list = &etr_devs; if (!of_property_read_u32(dev->of_node, "csr-atid-offset", &drvdata->atid_offset)) coresight_set_csr_ops(&csr_atid_ops); break; case TMC_CONFIG_TYPE_ETF: desc.type = CORESIGHT_DEV_TYPE_LINKSINK; desc.subtype.link_subtype = CORESIGHT_DEV_SUBTYPE_LINK_FIFO; desc.groups = coresight_tmc_etf_groups; desc.ops = &tmc_etf_cs_ops; dev_list = &etf_devs; break; default: pr_err("%s: Unsupported TMC config\n", desc.name); ret = -EINVAL; goto out; } if (of_property_read_string(dev->of_node, "coresight-name", &desc.name)) desc.name = coresight_alloc_device_name(dev_list, dev); if (!desc.name) { ret = -ENOMEM; goto out; } pdata = coresight_get_platform_data(dev); if (IS_ERR(pdata)) { ret = PTR_ERR(pdata); goto out; } adev->dev.platform_data = pdata; desc.pdata = pdata; drvdata->csdev = coresight_register(&desc); if (IS_ERR(drvdata->csdev)) { ret = PTR_ERR(drvdata->csdev); goto out; } drvdata->miscdev.name = desc.name; drvdata->miscdev.minor = MISC_DYNAMIC_MINOR; drvdata->miscdev.fops = &tmc_fops; ret = misc_register(&drvdata->miscdev); if (ret) coresight_unregister(drvdata->csdev); else pm_runtime_put(&adev->dev); out: return ret; } static const struct amba_id tmc_ids[] = { CS_AMBA_ID(0x000bb961), /* Coresight SoC 600 TMC-ETR/ETS */ CS_AMBA_ID_DATA(0x000bb9e8, (unsigned long)CORESIGHT_SOC_600_ETR_CAPS), /* Coresight SoC 600 TMC-ETB */ CS_AMBA_ID(0x000bb9e9), /* Coresight SoC 600 TMC-ETF */ CS_AMBA_ID(0x000bb9ea), { 0, 0}, }; static struct amba_driver tmc_driver = { .drv = { .name = "coresight-tmc", .owner = THIS_MODULE, .suppress_bind_attrs = true, }, .probe = tmc_probe, .id_table = tmc_ids, }; builtin_amba_driver(tmc_driver);