|  | /* | 
|  | * QEMU Freescale eTSEC Emulator | 
|  | * | 
|  | * Copyright (c) 2011-2013 AdaCore | 
|  | * | 
|  | * Permission is hereby granted, free of charge, to any person obtaining a copy | 
|  | * of this software and associated documentation files (the "Software"), to deal | 
|  | * in the Software without restriction, including without limitation the rights | 
|  | * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | 
|  | * copies of the Software, and to permit persons to whom the Software is | 
|  | * furnished to do so, subject to the following conditions: | 
|  | * | 
|  | * The above copyright notice and this permission notice shall be included in | 
|  | * all copies or substantial portions of the Software. | 
|  | * | 
|  | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | 
|  | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | 
|  | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | 
|  | * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | 
|  | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | 
|  | * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | 
|  | * THE SOFTWARE. | 
|  | */ | 
|  | #include "qemu/osdep.h" | 
|  | #include "net/checksum.h" | 
|  | #include "qemu/log.h" | 
|  | #include "etsec.h" | 
|  | #include "registers.h" | 
|  |  | 
|  | /* #define ETSEC_RING_DEBUG */ | 
|  | /* #define HEX_DUMP */ | 
|  | /* #define DEBUG_BD */ | 
|  |  | 
|  | #ifdef ETSEC_RING_DEBUG | 
|  | static const int debug_etsec = 1; | 
|  | #else | 
|  | static const int debug_etsec; | 
|  | #endif | 
|  |  | 
|  | #define RING_DEBUG(fmt, ...) do {              \ | 
|  | if (debug_etsec) {                            \ | 
|  | qemu_log(fmt , ## __VA_ARGS__);        \ | 
|  | }                                          \ | 
|  | } while (0) | 
|  |  | 
|  | #ifdef DEBUG_BD | 
|  |  | 
|  | static void print_tx_bd_flags(uint16_t flags) | 
|  | { | 
|  | qemu_log("      Ready: %d\n", !!(flags & BD_TX_READY)); | 
|  | qemu_log("      PAD/CRC: %d\n", !!(flags & BD_TX_PADCRC)); | 
|  | qemu_log("      Wrap: %d\n", !!(flags & BD_WRAP)); | 
|  | qemu_log("      Interrupt: %d\n", !!(flags & BD_INTERRUPT)); | 
|  | qemu_log("      Last in frame: %d\n", !!(flags & BD_LAST)); | 
|  | qemu_log("      Tx CRC: %d\n", !!(flags & BD_TX_TC)); | 
|  | qemu_log("      User-defined preamble / defer: %d\n", | 
|  | !!(flags & BD_TX_PREDEF)); | 
|  | qemu_log("      Huge frame enable / Late collision: %d\n", | 
|  | !!(flags & BD_TX_HFELC)); | 
|  | qemu_log("      Control frame / Retransmission Limit: %d\n", | 
|  | !!(flags & BD_TX_CFRL)); | 
|  | qemu_log("      Retry count: %d\n", | 
|  | (flags >> BD_TX_RC_OFFSET) & BD_TX_RC_MASK); | 
|  | qemu_log("      Underrun / TCP/IP off-load enable: %d\n", | 
|  | !!(flags & BD_TX_TOEUN)); | 
|  | qemu_log("      Truncation: %d\n", !!(flags & BD_TX_TR)); | 
|  | } | 
|  |  | 
|  | static void print_rx_bd_flags(uint16_t flags) | 
|  | { | 
|  | qemu_log("      Empty: %d\n", !!(flags & BD_RX_EMPTY)); | 
|  | qemu_log("      Receive software ownership: %d\n", !!(flags & BD_RX_RO1)); | 
|  | qemu_log("      Wrap: %d\n", !!(flags & BD_WRAP)); | 
|  | qemu_log("      Interrupt: %d\n", !!(flags & BD_INTERRUPT)); | 
|  | qemu_log("      Last in frame: %d\n", !!(flags & BD_LAST)); | 
|  | qemu_log("      First in frame: %d\n", !!(flags & BD_RX_FIRST)); | 
|  | qemu_log("      Miss: %d\n", !!(flags & BD_RX_MISS)); | 
|  | qemu_log("      Broadcast: %d\n", !!(flags & BD_RX_BROADCAST)); | 
|  | qemu_log("      Multicast: %d\n", !!(flags & BD_RX_MULTICAST)); | 
|  | qemu_log("      Rx frame length violation: %d\n", !!(flags & BD_RX_LG)); | 
|  | qemu_log("      Rx non-octet aligned frame: %d\n", !!(flags & BD_RX_NO)); | 
|  | qemu_log("      Short frame: %d\n", !!(flags & BD_RX_SH)); | 
|  | qemu_log("      Rx CRC Error: %d\n", !!(flags & BD_RX_CR)); | 
|  | qemu_log("      Overrun: %d\n", !!(flags & BD_RX_OV)); | 
|  | qemu_log("      Truncation: %d\n", !!(flags & BD_RX_TR)); | 
|  | } | 
|  |  | 
|  |  | 
|  | static void print_bd(eTSEC_rxtx_bd bd, int mode, uint32_t index) | 
|  | { | 
|  | qemu_log("eTSEC %s Data Buffer Descriptor (%u)\n", | 
|  | mode == eTSEC_TRANSMIT ? "Transmit" : "Receive", | 
|  | index); | 
|  | qemu_log("   Flags   : 0x%04x\n", bd.flags); | 
|  | if (mode == eTSEC_TRANSMIT) { | 
|  | print_tx_bd_flags(bd.flags); | 
|  | } else { | 
|  | print_rx_bd_flags(bd.flags); | 
|  | } | 
|  | qemu_log("   Length  : 0x%04x\n", bd.length); | 
|  | qemu_log("   Pointer : 0x%08x\n", bd.bufptr); | 
|  | } | 
|  |  | 
|  | #endif  /* DEBUG_BD */ | 
|  |  | 
|  | static void read_buffer_descriptor(eTSEC         *etsec, | 
|  | hwaddr         addr, | 
|  | eTSEC_rxtx_bd *bd) | 
|  | { | 
|  | assert(bd != NULL); | 
|  |  | 
|  | RING_DEBUG("READ Buffer Descriptor @ 0x" HWADDR_FMT_plx"\n", addr); | 
|  | cpu_physical_memory_read(addr, | 
|  | bd, | 
|  | sizeof(eTSEC_rxtx_bd)); | 
|  |  | 
|  | if (etsec->regs[DMACTRL].value & DMACTRL_LE) { | 
|  | bd->flags  = lduw_le_p(&bd->flags); | 
|  | bd->length = lduw_le_p(&bd->length); | 
|  | bd->bufptr = ldl_le_p(&bd->bufptr); | 
|  | } else { | 
|  | bd->flags  = lduw_be_p(&bd->flags); | 
|  | bd->length = lduw_be_p(&bd->length); | 
|  | bd->bufptr = ldl_be_p(&bd->bufptr); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void write_buffer_descriptor(eTSEC         *etsec, | 
|  | hwaddr         addr, | 
|  | eTSEC_rxtx_bd *bd) | 
|  | { | 
|  | assert(bd != NULL); | 
|  |  | 
|  | if (etsec->regs[DMACTRL].value & DMACTRL_LE) { | 
|  | stw_le_p(&bd->flags, bd->flags); | 
|  | stw_le_p(&bd->length, bd->length); | 
|  | stl_le_p(&bd->bufptr, bd->bufptr); | 
|  | } else { | 
|  | stw_be_p(&bd->flags, bd->flags); | 
|  | stw_be_p(&bd->length, bd->length); | 
|  | stl_be_p(&bd->bufptr, bd->bufptr); | 
|  | } | 
|  |  | 
|  | RING_DEBUG("Write Buffer Descriptor @ 0x" HWADDR_FMT_plx"\n", addr); | 
|  | cpu_physical_memory_write(addr, | 
|  | bd, | 
|  | sizeof(eTSEC_rxtx_bd)); | 
|  | } | 
|  |  | 
|  | static void ievent_set(eTSEC    *etsec, | 
|  | uint32_t  flags) | 
|  | { | 
|  | etsec->regs[IEVENT].value |= flags; | 
|  |  | 
|  | etsec_update_irq(etsec); | 
|  | } | 
|  |  | 
|  | static void tx_padding_and_crc(eTSEC *etsec, uint32_t min_frame_len) | 
|  | { | 
|  | int add = min_frame_len - etsec->tx_buffer_len; | 
|  |  | 
|  | /* Padding */ | 
|  | if (add > 0) { | 
|  | RING_DEBUG("pad:%u\n", add); | 
|  | etsec->tx_buffer = g_realloc(etsec->tx_buffer, | 
|  | etsec->tx_buffer_len + add); | 
|  |  | 
|  | memset(etsec->tx_buffer + etsec->tx_buffer_len, 0x0, add); | 
|  | etsec->tx_buffer_len += add; | 
|  | } | 
|  |  | 
|  | /* Never add CRC in QEMU */ | 
|  | } | 
|  |  | 
|  | static void process_tx_fcb(eTSEC *etsec) | 
|  | { | 
|  | uint8_t flags = (uint8_t)(*etsec->tx_buffer); | 
|  | /* L3 header offset from start of frame */ | 
|  | uint8_t l3_header_offset = (uint8_t)*(etsec->tx_buffer + 3); | 
|  | /* L4 header offset from start of L3 header */ | 
|  | uint8_t l4_header_offset = (uint8_t)*(etsec->tx_buffer + 2); | 
|  | /* L3 header */ | 
|  | uint8_t *l3_header = etsec->tx_buffer + 8 + l3_header_offset; | 
|  | /* L4 header */ | 
|  | uint8_t *l4_header = l3_header + l4_header_offset; | 
|  | int csum = 0; | 
|  |  | 
|  | /* if packet is IP4 and IP checksum is requested */ | 
|  | if (flags & FCB_TX_IP && flags & FCB_TX_CIP) { | 
|  | csum |= CSUM_IP; | 
|  | } | 
|  | /* TODO Check the correct usage of the PHCS field of the FCB in case the NPH | 
|  | * flag is on */ | 
|  |  | 
|  | /* if packet is IP4 and TCP or UDP */ | 
|  | if (flags & FCB_TX_IP && flags & FCB_TX_TUP) { | 
|  | /* if UDP */ | 
|  | if (flags & FCB_TX_UDP) { | 
|  | /* if checksum is requested */ | 
|  | if (flags & FCB_TX_CTU) { | 
|  | /* do UDP checksum */ | 
|  | csum |= CSUM_UDP; | 
|  | } else { | 
|  | /* set checksum field to 0 */ | 
|  | l4_header[6] = 0; | 
|  | l4_header[7] = 0; | 
|  | } | 
|  | } else if (flags & FCB_TX_CTU) { /* if TCP and checksum is requested */ | 
|  | /* do TCP checksum */ | 
|  | csum |= CSUM_TCP; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (csum) { | 
|  | net_checksum_calculate(etsec->tx_buffer + 8, | 
|  | etsec->tx_buffer_len - 8, csum); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void process_tx_bd(eTSEC         *etsec, | 
|  | eTSEC_rxtx_bd *bd) | 
|  | { | 
|  | uint8_t *tmp_buff = NULL; | 
|  | hwaddr tbdbth     = (hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32; | 
|  |  | 
|  | if (bd->length == 0) { | 
|  | /* ERROR */ | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (etsec->tx_buffer_len == 0) { | 
|  | /* It's the first BD */ | 
|  | etsec->first_bd = *bd; | 
|  | } | 
|  |  | 
|  | /* TODO: if TxBD[TOE/UN] skip the Tx Frame Control Block*/ | 
|  |  | 
|  | /* Load this Data Buffer */ | 
|  | etsec->tx_buffer = g_realloc(etsec->tx_buffer, | 
|  | etsec->tx_buffer_len + bd->length); | 
|  | tmp_buff = etsec->tx_buffer + etsec->tx_buffer_len; | 
|  | cpu_physical_memory_read(bd->bufptr + tbdbth, tmp_buff, bd->length); | 
|  |  | 
|  | /* Update buffer length */ | 
|  | etsec->tx_buffer_len += bd->length; | 
|  |  | 
|  |  | 
|  | if (etsec->tx_buffer_len != 0 && (bd->flags & BD_LAST)) { | 
|  | if (etsec->regs[MACCFG1].value & MACCFG1_TX_EN) { | 
|  | /* MAC Transmit enabled */ | 
|  |  | 
|  | /* Process offload Tx FCB */ | 
|  | if (etsec->first_bd.flags & BD_TX_TOEUN) { | 
|  | process_tx_fcb(etsec); | 
|  | } | 
|  |  | 
|  | if (etsec->first_bd.flags & BD_TX_PADCRC | 
|  | || etsec->regs[MACCFG2].value & MACCFG2_PADCRC) { | 
|  |  | 
|  | /* Padding and CRC (Padding implies CRC) */ | 
|  | tx_padding_and_crc(etsec, 60); | 
|  |  | 
|  | } else if (etsec->first_bd.flags & BD_TX_TC | 
|  | || etsec->regs[MACCFG2].value & MACCFG2_CRC_EN) { | 
|  |  | 
|  | /* Only CRC */ | 
|  | /* Never add CRC in QEMU */ | 
|  | } | 
|  |  | 
|  | #if defined(HEX_DUMP) | 
|  | qemu_log("eTSEC Send packet size:%d\n", etsec->tx_buffer_len); | 
|  | qemu_hexdump(stderr, "", etsec->tx_buffer, etsec->tx_buffer_len); | 
|  | #endif  /* ETSEC_RING_DEBUG */ | 
|  |  | 
|  | if (etsec->first_bd.flags & BD_TX_TOEUN) { | 
|  | qemu_send_packet(qemu_get_queue(etsec->nic), | 
|  | etsec->tx_buffer + 8, | 
|  | etsec->tx_buffer_len - 8); | 
|  | } else { | 
|  | qemu_send_packet(qemu_get_queue(etsec->nic), | 
|  | etsec->tx_buffer, | 
|  | etsec->tx_buffer_len); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | etsec->tx_buffer_len = 0; | 
|  |  | 
|  | if (bd->flags & BD_INTERRUPT) { | 
|  | ievent_set(etsec, IEVENT_TXF); | 
|  | } | 
|  | } else { | 
|  | if (bd->flags & BD_INTERRUPT) { | 
|  | ievent_set(etsec, IEVENT_TXB); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Update DB flags */ | 
|  |  | 
|  | /* Clear Ready */ | 
|  | bd->flags &= ~BD_TX_READY; | 
|  |  | 
|  | /* Clear Defer */ | 
|  | bd->flags &= ~BD_TX_PREDEF; | 
|  |  | 
|  | /* Clear Late Collision */ | 
|  | bd->flags &= ~BD_TX_HFELC; | 
|  |  | 
|  | /* Clear Retransmission Limit */ | 
|  | bd->flags &= ~BD_TX_CFRL; | 
|  |  | 
|  | /* Clear Retry Count */ | 
|  | bd->flags &= ~(BD_TX_RC_MASK << BD_TX_RC_OFFSET); | 
|  |  | 
|  | /* Clear Underrun */ | 
|  | bd->flags &= ~BD_TX_TOEUN; | 
|  |  | 
|  | /* Clear Truncation */ | 
|  | bd->flags &= ~BD_TX_TR; | 
|  | } | 
|  |  | 
|  | void etsec_walk_tx_ring(eTSEC *etsec, int ring_nbr) | 
|  | { | 
|  | hwaddr        ring_base = 0; | 
|  | hwaddr        bd_addr   = 0; | 
|  | eTSEC_rxtx_bd bd; | 
|  | uint16_t      bd_flags; | 
|  |  | 
|  | if (!(etsec->regs[MACCFG1].value & MACCFG1_TX_EN)) { | 
|  | RING_DEBUG("%s: MAC Transmit not enabled\n", __func__); | 
|  | return; | 
|  | } | 
|  |  | 
|  | ring_base = (hwaddr)(etsec->regs[TBASEH].value & 0xF) << 32; | 
|  | ring_base += etsec->regs[TBASE0 + ring_nbr].value & ~0x7; | 
|  | bd_addr    = etsec->regs[TBPTR0 + ring_nbr].value & ~0x7; | 
|  |  | 
|  | do { | 
|  | read_buffer_descriptor(etsec, bd_addr, &bd); | 
|  |  | 
|  | #ifdef DEBUG_BD | 
|  | print_bd(bd, | 
|  | eTSEC_TRANSMIT, | 
|  | (bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd)); | 
|  |  | 
|  | #endif  /* DEBUG_BD */ | 
|  |  | 
|  | /* Save flags before BD update */ | 
|  | bd_flags = bd.flags; | 
|  |  | 
|  | if (!(bd_flags & BD_TX_READY)) { | 
|  | break; | 
|  | } | 
|  |  | 
|  | process_tx_bd(etsec, &bd); | 
|  | /* Write back BD after update */ | 
|  | write_buffer_descriptor(etsec, bd_addr, &bd); | 
|  |  | 
|  | /* Wrap or next BD */ | 
|  | if (bd_flags & BD_WRAP) { | 
|  | bd_addr = ring_base; | 
|  | } else { | 
|  | bd_addr += sizeof(eTSEC_rxtx_bd); | 
|  | } | 
|  | } while (TRUE); | 
|  |  | 
|  | /* Save the Buffer Descriptor Pointers to last bd that was not | 
|  | * successfully closed */ | 
|  | etsec->regs[TBPTR0 + ring_nbr].value = bd_addr; | 
|  |  | 
|  | /* Set transmit halt THLTx */ | 
|  | etsec->regs[TSTAT].value |= 1 << (31 - ring_nbr); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * rx_init_frame() ensures we never do more padding than this | 
|  | * (checksum plus minimum data packet size) | 
|  | */ | 
|  | #define MAX_RX_PADDING 64 | 
|  |  | 
|  | static void fill_rx_bd(eTSEC          *etsec, | 
|  | eTSEC_rxtx_bd  *bd, | 
|  | const uint8_t **buf, | 
|  | size_t         *size) | 
|  | { | 
|  | uint16_t to_write; | 
|  | hwaddr   bufptr = bd->bufptr + | 
|  | ((hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32); | 
|  | uint8_t  padd[MAX_RX_PADDING]; | 
|  | uint8_t  rem; | 
|  |  | 
|  | assert(etsec->rx_padding <= MAX_RX_PADDING); | 
|  |  | 
|  | RING_DEBUG("eTSEC fill Rx buffer @ 0x%016" HWADDR_PRIx | 
|  | " size:%zu(padding + crc:%u) + fcb:%u\n", | 
|  | bufptr, *size, etsec->rx_padding, etsec->rx_fcb_size); | 
|  |  | 
|  | bd->length = 0; | 
|  |  | 
|  | /* This operation will only write FCB */ | 
|  | if (etsec->rx_fcb_size != 0) { | 
|  |  | 
|  | cpu_physical_memory_write(bufptr, etsec->rx_fcb, etsec->rx_fcb_size); | 
|  |  | 
|  | bufptr             += etsec->rx_fcb_size; | 
|  | bd->length         += etsec->rx_fcb_size; | 
|  | etsec->rx_fcb_size  = 0; | 
|  |  | 
|  | } | 
|  |  | 
|  | /* We remove padding from the computation of to_write because it is not | 
|  | * allocated in the buffer. | 
|  | */ | 
|  | to_write = MIN(*size - etsec->rx_padding, | 
|  | etsec->regs[MRBLR].value - etsec->rx_fcb_size); | 
|  |  | 
|  | /* This operation can only write packet data and no padding */ | 
|  | if (to_write > 0) { | 
|  | cpu_physical_memory_write(bufptr, *buf, to_write); | 
|  |  | 
|  | *buf   += to_write; | 
|  | bufptr += to_write; | 
|  | *size  -= to_write; | 
|  |  | 
|  | bd->flags  &= ~BD_RX_EMPTY; | 
|  | bd->length += to_write; | 
|  | } | 
|  |  | 
|  | if (*size == etsec->rx_padding) { | 
|  | /* The remaining bytes are only for padding which is not actually | 
|  | * allocated in the data buffer. | 
|  | */ | 
|  |  | 
|  | rem = MIN(etsec->regs[MRBLR].value - bd->length, etsec->rx_padding); | 
|  |  | 
|  | if (rem > 0) { | 
|  | memset(padd, 0x0, rem); | 
|  | etsec->rx_padding -= rem; | 
|  | *size             -= rem; | 
|  | bd->length        += rem; | 
|  | cpu_physical_memory_write(bufptr, padd, rem); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void rx_init_frame(eTSEC *etsec, const uint8_t *buf, size_t size) | 
|  | { | 
|  | uint32_t fcb_size = 0; | 
|  | uint8_t  prsdep   = (etsec->regs[RCTRL].value >> RCTRL_PRSDEP_OFFSET) | 
|  | & RCTRL_PRSDEP_MASK; | 
|  |  | 
|  | if (prsdep != 0) { | 
|  | /* Prepend FCB (FCB size + RCTRL[PAL]) */ | 
|  | fcb_size = 8 + ((etsec->regs[RCTRL].value >> 16) & 0x1F); | 
|  |  | 
|  | etsec->rx_fcb_size = fcb_size; | 
|  |  | 
|  | /* TODO: fill_FCB(etsec); */ | 
|  | memset(etsec->rx_fcb, 0x0, sizeof(etsec->rx_fcb)); | 
|  |  | 
|  | } else { | 
|  | etsec->rx_fcb_size = 0; | 
|  | } | 
|  |  | 
|  | g_free(etsec->rx_buffer); | 
|  |  | 
|  | /* Do not copy the frame for now */ | 
|  | etsec->rx_buffer     = (uint8_t *)buf; | 
|  | etsec->rx_buffer_len = size; | 
|  |  | 
|  | /* CRC padding (We don't have to compute the CRC) */ | 
|  | etsec->rx_padding = 4; | 
|  |  | 
|  | /* | 
|  | * Ensure that payload length + CRC length is at least 802.3 | 
|  | * minimum MTU size bytes long (64) | 
|  | */ | 
|  | if (etsec->rx_buffer_len < 60) { | 
|  | etsec->rx_padding += 60 - etsec->rx_buffer_len; | 
|  | } | 
|  |  | 
|  | etsec->rx_first_in_frame = 1; | 
|  | etsec->rx_remaining_data = etsec->rx_buffer_len; | 
|  | RING_DEBUG("%s: rx_buffer_len:%u rx_padding+crc:%u\n", __func__, | 
|  | etsec->rx_buffer_len, etsec->rx_padding); | 
|  | } | 
|  |  | 
|  | ssize_t etsec_rx_ring_write(eTSEC *etsec, const uint8_t *buf, size_t size) | 
|  | { | 
|  | int ring_nbr = 0;           /* Always use ring0 (no filer) */ | 
|  |  | 
|  | if (etsec->rx_buffer_len != 0) { | 
|  | RING_DEBUG("%s: We can't receive now," | 
|  | " a buffer is already in the pipe\n", __func__); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (etsec->regs[RSTAT].value & 1 << (23 - ring_nbr)) { | 
|  | RING_DEBUG("%s: The ring is halted\n", __func__); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (etsec->regs[DMACTRL].value & DMACTRL_GRS) { | 
|  | RING_DEBUG("%s: Graceful receive stop\n", __func__); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (!(etsec->regs[MACCFG1].value & MACCFG1_RX_EN)) { | 
|  | RING_DEBUG("%s: MAC Receive not enabled\n", __func__); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (!(etsec->regs[RCTRL].value & RCTRL_RSF) && (size < 60)) { | 
|  | /* CRC is not in the packet yet, so short frame is below 60 bytes */ | 
|  | RING_DEBUG("%s: Drop short frame\n", __func__); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | rx_init_frame(etsec, buf, size); | 
|  |  | 
|  | etsec_walk_rx_ring(etsec, ring_nbr); | 
|  |  | 
|  | return size; | 
|  | } | 
|  |  | 
|  | void etsec_walk_rx_ring(eTSEC *etsec, int ring_nbr) | 
|  | { | 
|  | hwaddr         ring_base     = 0; | 
|  | hwaddr         bd_addr       = 0; | 
|  | hwaddr         start_bd_addr = 0; | 
|  | eTSEC_rxtx_bd  bd; | 
|  | uint16_t       bd_flags; | 
|  | size_t         remaining_data; | 
|  | const uint8_t *buf; | 
|  | uint8_t       *tmp_buf; | 
|  | size_t         size; | 
|  |  | 
|  | if (etsec->rx_buffer_len == 0) { | 
|  | /* No frame to send */ | 
|  | RING_DEBUG("No frame to send\n"); | 
|  | return; | 
|  | } | 
|  |  | 
|  | remaining_data = etsec->rx_remaining_data + etsec->rx_padding; | 
|  | buf            = etsec->rx_buffer | 
|  | + (etsec->rx_buffer_len - etsec->rx_remaining_data); | 
|  | size           = etsec->rx_buffer_len + etsec->rx_padding; | 
|  |  | 
|  | ring_base = (hwaddr)(etsec->regs[RBASEH].value & 0xF) << 32; | 
|  | ring_base += etsec->regs[RBASE0 + ring_nbr].value & ~0x7; | 
|  | start_bd_addr  = bd_addr = etsec->regs[RBPTR0 + ring_nbr].value & ~0x7; | 
|  |  | 
|  | do { | 
|  | read_buffer_descriptor(etsec, bd_addr, &bd); | 
|  |  | 
|  | #ifdef DEBUG_BD | 
|  | print_bd(bd, | 
|  | eTSEC_RECEIVE, | 
|  | (bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd)); | 
|  |  | 
|  | #endif  /* DEBUG_BD */ | 
|  |  | 
|  | /* Save flags before BD update */ | 
|  | bd_flags = bd.flags; | 
|  |  | 
|  | if (bd_flags & BD_RX_EMPTY) { | 
|  | fill_rx_bd(etsec, &bd, &buf, &remaining_data); | 
|  |  | 
|  | if (etsec->rx_first_in_frame) { | 
|  | bd.flags |= BD_RX_FIRST; | 
|  | etsec->rx_first_in_frame = 0; | 
|  | etsec->rx_first_bd = bd; | 
|  | } | 
|  |  | 
|  | /* Last in frame */ | 
|  | if (remaining_data == 0) { | 
|  |  | 
|  | /* Clear flags */ | 
|  |  | 
|  | bd.flags &= ~0x7ff; | 
|  |  | 
|  | bd.flags |= BD_LAST; | 
|  |  | 
|  | /* NOTE: non-octet aligned frame is impossible in qemu */ | 
|  |  | 
|  | if (size >= etsec->regs[MAXFRM].value) { | 
|  | /* frame length violation */ | 
|  | qemu_log("%s frame length violation: size:%zu MAXFRM:%d\n", | 
|  | __func__, size, etsec->regs[MAXFRM].value); | 
|  |  | 
|  | bd.flags |= BD_RX_LG; | 
|  | } | 
|  |  | 
|  | if (size  < 64) { | 
|  | /* Short frame */ | 
|  | bd.flags |= BD_RX_SH; | 
|  | } | 
|  |  | 
|  | /* TODO: Broadcast and Multicast */ | 
|  |  | 
|  | if (bd.flags & BD_INTERRUPT) { | 
|  | /* Set RXFx */ | 
|  | etsec->regs[RSTAT].value |= 1 << (7 - ring_nbr); | 
|  |  | 
|  | /* Set IEVENT */ | 
|  | ievent_set(etsec, IEVENT_RXF); | 
|  | } | 
|  |  | 
|  | } else { | 
|  | if (bd.flags & BD_INTERRUPT) { | 
|  | /* Set IEVENT */ | 
|  | ievent_set(etsec, IEVENT_RXB); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Write back BD after update */ | 
|  | write_buffer_descriptor(etsec, bd_addr, &bd); | 
|  | } | 
|  |  | 
|  | /* Wrap or next BD */ | 
|  | if (bd_flags & BD_WRAP) { | 
|  | bd_addr = ring_base; | 
|  | } else { | 
|  | bd_addr += sizeof(eTSEC_rxtx_bd); | 
|  | } | 
|  | } while (remaining_data != 0 | 
|  | && (bd_flags & BD_RX_EMPTY) | 
|  | && bd_addr != start_bd_addr); | 
|  |  | 
|  | /* Reset ring ptr */ | 
|  | etsec->regs[RBPTR0 + ring_nbr].value = bd_addr; | 
|  |  | 
|  | /* The frame is too large to fit in the Rx ring */ | 
|  | if (remaining_data > 0) { | 
|  |  | 
|  | /* Set RSTAT[QHLTx] */ | 
|  | etsec->regs[RSTAT].value |= 1 << (23 - ring_nbr); | 
|  |  | 
|  | /* Save remaining data to send the end of the frame when the ring will | 
|  | * be restarted | 
|  | */ | 
|  | etsec->rx_remaining_data = remaining_data; | 
|  |  | 
|  | /* Copy the frame */ | 
|  | tmp_buf = g_malloc(size); | 
|  | memcpy(tmp_buf, etsec->rx_buffer, size); | 
|  | etsec->rx_buffer = tmp_buf; | 
|  |  | 
|  | RING_DEBUG("no empty RxBD available any more\n"); | 
|  | } else { | 
|  | etsec->rx_buffer_len = 0; | 
|  | etsec->rx_buffer     = NULL; | 
|  | if (etsec->need_flush) { | 
|  | qemu_flush_queued_packets(qemu_get_queue(etsec->nic)); | 
|  | } | 
|  | } | 
|  |  | 
|  | RING_DEBUG("eTSEC End of ring_write: remaining_data:%zu\n", remaining_data); | 
|  | } |