| /* |
| * SiFive Platform DMA emulation |
| * |
| * Copyright (c) 2020 Wind River Systems, Inc. |
| * |
| * Author: |
| * Bin Meng <bin.meng@windriver.com> |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License as |
| * published by the Free Software Foundation; either version 2 or |
| * (at your option) version 3 of the License. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License along |
| * with this program; if not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/bitops.h" |
| #include "qemu/log.h" |
| #include "qapi/error.h" |
| #include "hw/irq.h" |
| #include "hw/qdev-properties.h" |
| #include "hw/sysbus.h" |
| #include "migration/vmstate.h" |
| #include "sysemu/dma.h" |
| #include "hw/dma/sifive_pdma.h" |
| |
| #define DMA_CONTROL 0x000 |
| #define CONTROL_CLAIM BIT(0) |
| #define CONTROL_RUN BIT(1) |
| #define CONTROL_DONE_IE BIT(14) |
| #define CONTROL_ERR_IE BIT(15) |
| #define CONTROL_DONE BIT(30) |
| #define CONTROL_ERR BIT(31) |
| |
| #define DMA_NEXT_CONFIG 0x004 |
| #define CONFIG_REPEAT BIT(2) |
| #define CONFIG_ORDER BIT(3) |
| #define CONFIG_WRSZ_SHIFT 24 |
| #define CONFIG_RDSZ_SHIFT 28 |
| #define CONFIG_SZ_MASK 0xf |
| |
| #define DMA_NEXT_BYTES 0x008 |
| #define DMA_NEXT_DST 0x010 |
| #define DMA_NEXT_SRC 0x018 |
| #define DMA_EXEC_CONFIG 0x104 |
| #define DMA_EXEC_BYTES 0x108 |
| #define DMA_EXEC_DST 0x110 |
| #define DMA_EXEC_SRC 0x118 |
| |
| /* |
| * FU540/FU740 docs are incorrect with NextConfig.wsize/rsize reset values. |
| * The reset values tested on Unleashed/Unmatched boards are 6 instead of 0. |
| */ |
| #define CONFIG_WRSZ_DEFAULT 6 |
| #define CONFIG_RDSZ_DEFAULT 6 |
| |
| enum dma_chan_state { |
| DMA_CHAN_STATE_IDLE, |
| DMA_CHAN_STATE_STARTED, |
| DMA_CHAN_STATE_ERROR, |
| DMA_CHAN_STATE_DONE |
| }; |
| |
| static void sifive_pdma_run(SiFivePDMAState *s, int ch) |
| { |
| uint64_t bytes = s->chan[ch].next_bytes; |
| uint64_t dst = s->chan[ch].next_dst; |
| uint64_t src = s->chan[ch].next_src; |
| uint32_t config = s->chan[ch].next_config; |
| int wsize, rsize, size, remainder; |
| uint8_t buf[64]; |
| int n; |
| |
| /* do nothing if bytes to transfer is zero */ |
| if (!bytes) { |
| goto done; |
| } |
| |
| /* |
| * The manual does not describe how the hardware behaviors when |
| * config.wsize and config.rsize are given different values. |
| * A common case is memory to memory DMA, and in this case they |
| * are normally the same. Abort if this expectation fails. |
| */ |
| wsize = (config >> CONFIG_WRSZ_SHIFT) & CONFIG_SZ_MASK; |
| rsize = (config >> CONFIG_RDSZ_SHIFT) & CONFIG_SZ_MASK; |
| if (wsize != rsize) { |
| goto error; |
| } |
| |
| /* |
| * Calculate the transaction size |
| * |
| * size field is base 2 logarithm of DMA transaction size, |
| * but there is an upper limit of 64 bytes per transaction. |
| */ |
| size = wsize; |
| if (size > 6) { |
| size = 6; |
| } |
| size = 1 << size; |
| remainder = bytes % size; |
| |
| /* indicate a DMA transfer is started */ |
| s->chan[ch].state = DMA_CHAN_STATE_STARTED; |
| s->chan[ch].control &= ~CONTROL_DONE; |
| s->chan[ch].control &= ~CONTROL_ERR; |
| |
| /* load the next_ registers into their exec_ counterparts */ |
| s->chan[ch].exec_config = config; |
| s->chan[ch].exec_bytes = bytes; |
| s->chan[ch].exec_dst = dst; |
| s->chan[ch].exec_src = src; |
| |
| for (n = 0; n < bytes / size; n++) { |
| cpu_physical_memory_read(s->chan[ch].exec_src, buf, size); |
| cpu_physical_memory_write(s->chan[ch].exec_dst, buf, size); |
| s->chan[ch].exec_src += size; |
| s->chan[ch].exec_dst += size; |
| s->chan[ch].exec_bytes -= size; |
| } |
| |
| if (remainder) { |
| cpu_physical_memory_read(s->chan[ch].exec_src, buf, remainder); |
| cpu_physical_memory_write(s->chan[ch].exec_dst, buf, remainder); |
| s->chan[ch].exec_src += remainder; |
| s->chan[ch].exec_dst += remainder; |
| s->chan[ch].exec_bytes -= remainder; |
| } |
| |
| /* reload exec_ registers if repeat is required */ |
| if (s->chan[ch].next_config & CONFIG_REPEAT) { |
| s->chan[ch].exec_bytes = bytes; |
| s->chan[ch].exec_dst = dst; |
| s->chan[ch].exec_src = src; |
| } |
| |
| done: |
| /* indicate a DMA transfer is done */ |
| s->chan[ch].state = DMA_CHAN_STATE_DONE; |
| s->chan[ch].control &= ~CONTROL_RUN; |
| s->chan[ch].control |= CONTROL_DONE; |
| return; |
| |
| error: |
| s->chan[ch].state = DMA_CHAN_STATE_ERROR; |
| s->chan[ch].control |= CONTROL_ERR; |
| return; |
| } |
| |
| static inline void sifive_pdma_update_irq(SiFivePDMAState *s, int ch) |
| { |
| bool done_ie, err_ie; |
| |
| done_ie = !!(s->chan[ch].control & CONTROL_DONE_IE); |
| err_ie = !!(s->chan[ch].control & CONTROL_ERR_IE); |
| |
| if (done_ie && (s->chan[ch].control & CONTROL_DONE)) { |
| qemu_irq_raise(s->irq[ch * 2]); |
| } else { |
| qemu_irq_lower(s->irq[ch * 2]); |
| } |
| |
| if (err_ie && (s->chan[ch].control & CONTROL_ERR)) { |
| qemu_irq_raise(s->irq[ch * 2 + 1]); |
| } else { |
| qemu_irq_lower(s->irq[ch * 2 + 1]); |
| } |
| |
| s->chan[ch].state = DMA_CHAN_STATE_IDLE; |
| } |
| |
| static uint64_t sifive_pdma_readq(SiFivePDMAState *s, int ch, hwaddr offset) |
| { |
| uint64_t val = 0; |
| |
| offset &= 0xfff; |
| switch (offset) { |
| case DMA_NEXT_BYTES: |
| val = s->chan[ch].next_bytes; |
| break; |
| case DMA_NEXT_DST: |
| val = s->chan[ch].next_dst; |
| break; |
| case DMA_NEXT_SRC: |
| val = s->chan[ch].next_src; |
| break; |
| case DMA_EXEC_BYTES: |
| val = s->chan[ch].exec_bytes; |
| break; |
| case DMA_EXEC_DST: |
| val = s->chan[ch].exec_dst; |
| break; |
| case DMA_EXEC_SRC: |
| val = s->chan[ch].exec_src; |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "%s: Unexpected 64-bit access to 0x%" HWADDR_PRIX "\n", |
| __func__, offset); |
| break; |
| } |
| |
| return val; |
| } |
| |
| static uint32_t sifive_pdma_readl(SiFivePDMAState *s, int ch, hwaddr offset) |
| { |
| uint32_t val = 0; |
| |
| offset &= 0xfff; |
| switch (offset) { |
| case DMA_CONTROL: |
| val = s->chan[ch].control; |
| break; |
| case DMA_NEXT_CONFIG: |
| val = s->chan[ch].next_config; |
| break; |
| case DMA_NEXT_BYTES: |
| val = extract64(s->chan[ch].next_bytes, 0, 32); |
| break; |
| case DMA_NEXT_BYTES + 4: |
| val = extract64(s->chan[ch].next_bytes, 32, 32); |
| break; |
| case DMA_NEXT_DST: |
| val = extract64(s->chan[ch].next_dst, 0, 32); |
| break; |
| case DMA_NEXT_DST + 4: |
| val = extract64(s->chan[ch].next_dst, 32, 32); |
| break; |
| case DMA_NEXT_SRC: |
| val = extract64(s->chan[ch].next_src, 0, 32); |
| break; |
| case DMA_NEXT_SRC + 4: |
| val = extract64(s->chan[ch].next_src, 32, 32); |
| break; |
| case DMA_EXEC_CONFIG: |
| val = s->chan[ch].exec_config; |
| break; |
| case DMA_EXEC_BYTES: |
| val = extract64(s->chan[ch].exec_bytes, 0, 32); |
| break; |
| case DMA_EXEC_BYTES + 4: |
| val = extract64(s->chan[ch].exec_bytes, 32, 32); |
| break; |
| case DMA_EXEC_DST: |
| val = extract64(s->chan[ch].exec_dst, 0, 32); |
| break; |
| case DMA_EXEC_DST + 4: |
| val = extract64(s->chan[ch].exec_dst, 32, 32); |
| break; |
| case DMA_EXEC_SRC: |
| val = extract64(s->chan[ch].exec_src, 0, 32); |
| break; |
| case DMA_EXEC_SRC + 4: |
| val = extract64(s->chan[ch].exec_src, 32, 32); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "%s: Unexpected 32-bit access to 0x%" HWADDR_PRIX "\n", |
| __func__, offset); |
| break; |
| } |
| |
| return val; |
| } |
| |
| static uint64_t sifive_pdma_read(void *opaque, hwaddr offset, unsigned size) |
| { |
| SiFivePDMAState *s = opaque; |
| int ch = SIFIVE_PDMA_CHAN_NO(offset); |
| uint64_t val = 0; |
| |
| if (ch >= SIFIVE_PDMA_CHANS) { |
| qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid channel no %d\n", |
| __func__, ch); |
| return 0; |
| } |
| |
| switch (size) { |
| case 8: |
| val = sifive_pdma_readq(s, ch, offset); |
| break; |
| case 4: |
| val = sifive_pdma_readl(s, ch, offset); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid read size %u to PDMA\n", |
| __func__, size); |
| return 0; |
| } |
| |
| return val; |
| } |
| |
| static void sifive_pdma_writeq(SiFivePDMAState *s, int ch, |
| hwaddr offset, uint64_t value) |
| { |
| offset &= 0xfff; |
| switch (offset) { |
| case DMA_NEXT_BYTES: |
| s->chan[ch].next_bytes = value; |
| break; |
| case DMA_NEXT_DST: |
| s->chan[ch].next_dst = value; |
| break; |
| case DMA_NEXT_SRC: |
| s->chan[ch].next_src = value; |
| break; |
| case DMA_EXEC_BYTES: |
| case DMA_EXEC_DST: |
| case DMA_EXEC_SRC: |
| /* these are read-only registers */ |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "%s: Unexpected 64-bit access to 0x%" HWADDR_PRIX "\n", |
| __func__, offset); |
| break; |
| } |
| } |
| |
| static void sifive_pdma_writel(SiFivePDMAState *s, int ch, |
| hwaddr offset, uint32_t value) |
| { |
| bool claimed, run; |
| |
| offset &= 0xfff; |
| switch (offset) { |
| case DMA_CONTROL: |
| claimed = !!(s->chan[ch].control & CONTROL_CLAIM); |
| run = !!(s->chan[ch].control & CONTROL_RUN); |
| |
| if (!claimed && (value & CONTROL_CLAIM)) { |
| /* reset Next* registers */ |
| s->chan[ch].next_config = (CONFIG_RDSZ_DEFAULT << CONFIG_RDSZ_SHIFT) | |
| (CONFIG_WRSZ_DEFAULT << CONFIG_WRSZ_SHIFT); |
| s->chan[ch].next_bytes = 0; |
| s->chan[ch].next_dst = 0; |
| s->chan[ch].next_src = 0; |
| } |
| |
| /* claim bit can only be cleared when run is low */ |
| if (run && !(value & CONTROL_CLAIM)) { |
| value |= CONTROL_CLAIM; |
| } |
| |
| s->chan[ch].control = value; |
| |
| /* |
| * If channel was not claimed before run bit is set, |
| * or if the channel is disclaimed when run was low, |
| * DMA won't run. |
| */ |
| if (!claimed || (!run && !(value & CONTROL_CLAIM))) { |
| s->chan[ch].control &= ~CONTROL_RUN; |
| return; |
| } |
| |
| if (value & CONTROL_RUN) { |
| sifive_pdma_run(s, ch); |
| } |
| |
| sifive_pdma_update_irq(s, ch); |
| break; |
| case DMA_NEXT_CONFIG: |
| s->chan[ch].next_config = value; |
| break; |
| case DMA_NEXT_BYTES: |
| s->chan[ch].next_bytes = |
| deposit64(s->chan[ch].next_bytes, 0, 32, value); |
| break; |
| case DMA_NEXT_BYTES + 4: |
| s->chan[ch].next_bytes = |
| deposit64(s->chan[ch].next_bytes, 32, 32, value); |
| break; |
| case DMA_NEXT_DST: |
| s->chan[ch].next_dst = deposit64(s->chan[ch].next_dst, 0, 32, value); |
| break; |
| case DMA_NEXT_DST + 4: |
| s->chan[ch].next_dst = deposit64(s->chan[ch].next_dst, 32, 32, value); |
| break; |
| case DMA_NEXT_SRC: |
| s->chan[ch].next_src = deposit64(s->chan[ch].next_src, 0, 32, value); |
| break; |
| case DMA_NEXT_SRC + 4: |
| s->chan[ch].next_src = deposit64(s->chan[ch].next_src, 32, 32, value); |
| break; |
| case DMA_EXEC_CONFIG: |
| case DMA_EXEC_BYTES: |
| case DMA_EXEC_BYTES + 4: |
| case DMA_EXEC_DST: |
| case DMA_EXEC_DST + 4: |
| case DMA_EXEC_SRC: |
| case DMA_EXEC_SRC + 4: |
| /* these are read-only registers */ |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "%s: Unexpected 32-bit access to 0x%" HWADDR_PRIX "\n", |
| __func__, offset); |
| break; |
| } |
| } |
| |
| static void sifive_pdma_write(void *opaque, hwaddr offset, |
| uint64_t value, unsigned size) |
| { |
| SiFivePDMAState *s = opaque; |
| int ch = SIFIVE_PDMA_CHAN_NO(offset); |
| |
| if (ch >= SIFIVE_PDMA_CHANS) { |
| qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid channel no %d\n", |
| __func__, ch); |
| return; |
| } |
| |
| switch (size) { |
| case 8: |
| sifive_pdma_writeq(s, ch, offset, value); |
| break; |
| case 4: |
| sifive_pdma_writel(s, ch, offset, (uint32_t) value); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid write size %u to PDMA\n", |
| __func__, size); |
| break; |
| } |
| } |
| |
| static const MemoryRegionOps sifive_pdma_ops = { |
| .read = sifive_pdma_read, |
| .write = sifive_pdma_write, |
| .endianness = DEVICE_LITTLE_ENDIAN, |
| /* there are 32-bit and 64-bit wide registers */ |
| .impl = { |
| .min_access_size = 4, |
| .max_access_size = 8, |
| }, |
| .valid = { |
| .min_access_size = 4, |
| .max_access_size = 8, |
| } |
| }; |
| |
| static void sifive_pdma_realize(DeviceState *dev, Error **errp) |
| { |
| SiFivePDMAState *s = SIFIVE_PDMA(dev); |
| int i; |
| |
| memory_region_init_io(&s->iomem, OBJECT(dev), &sifive_pdma_ops, s, |
| TYPE_SIFIVE_PDMA, SIFIVE_PDMA_REG_SIZE); |
| sysbus_init_mmio(SYS_BUS_DEVICE(dev), &s->iomem); |
| |
| for (i = 0; i < SIFIVE_PDMA_IRQS; i++) { |
| sysbus_init_irq(SYS_BUS_DEVICE(dev), &s->irq[i]); |
| } |
| } |
| |
| static void sifive_pdma_class_init(ObjectClass *klass, void *data) |
| { |
| DeviceClass *dc = DEVICE_CLASS(klass); |
| |
| dc->desc = "SiFive Platform DMA controller"; |
| dc->realize = sifive_pdma_realize; |
| } |
| |
| static const TypeInfo sifive_pdma_info = { |
| .name = TYPE_SIFIVE_PDMA, |
| .parent = TYPE_SYS_BUS_DEVICE, |
| .instance_size = sizeof(SiFivePDMAState), |
| .class_init = sifive_pdma_class_init, |
| }; |
| |
| static void sifive_pdma_register_types(void) |
| { |
| type_register_static(&sifive_pdma_info); |
| } |
| |
| type_init(sifive_pdma_register_types) |