qemu/hw/net/fsl_etsec/rings.c
<<
>>
Prefs
   1/*
   2 * QEMU Freescale eTSEC Emulator
   3 *
   4 * Copyright (c) 2011-2013 AdaCore
   5 *
   6 * Permission is hereby granted, free of charge, to any person obtaining a copy
   7 * of this software and associated documentation files (the "Software"), to deal
   8 * in the Software without restriction, including without limitation the rights
   9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
  10 * copies of the Software, and to permit persons to whom the Software is
  11 * furnished to do so, subject to the following conditions:
  12 *
  13 * The above copyright notice and this permission notice shall be included in
  14 * all copies or substantial portions of the Software.
  15 *
  16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
  21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
  22 * THE SOFTWARE.
  23 */
  24#include "qemu/osdep.h"
  25#include "qemu-common.h"
  26#include "net/checksum.h"
  27#include "qemu/log.h"
  28#include "etsec.h"
  29#include "registers.h"
  30
  31/* #define ETSEC_RING_DEBUG */
  32/* #define HEX_DUMP */
  33/* #define DEBUG_BD */
  34
  35#ifdef ETSEC_RING_DEBUG
  36static const int debug_etsec = 1;
  37#else
  38static const int debug_etsec;
  39#endif
  40
  41#define RING_DEBUG(fmt, ...) do {              \
  42 if (debug_etsec) {                            \
  43        qemu_log(fmt , ## __VA_ARGS__);        \
  44    }                                          \
  45    } while (0)
  46
  47#ifdef DEBUG_BD
  48
  49static void print_tx_bd_flags(uint16_t flags)
  50{
  51    qemu_log("      Ready: %d\n", !!(flags & BD_TX_READY));
  52    qemu_log("      PAD/CRC: %d\n", !!(flags & BD_TX_PADCRC));
  53    qemu_log("      Wrap: %d\n", !!(flags & BD_WRAP));
  54    qemu_log("      Interrupt: %d\n", !!(flags & BD_INTERRUPT));
  55    qemu_log("      Last in frame: %d\n", !!(flags & BD_LAST));
  56    qemu_log("      Tx CRC: %d\n", !!(flags & BD_TX_TC));
  57    qemu_log("      User-defined preamble / defer: %d\n",
  58           !!(flags & BD_TX_PREDEF));
  59    qemu_log("      Huge frame enable / Late collision: %d\n",
  60           !!(flags & BD_TX_HFELC));
  61    qemu_log("      Control frame / Retransmission Limit: %d\n",
  62           !!(flags & BD_TX_CFRL));
  63    qemu_log("      Retry count: %d\n",
  64           (flags >> BD_TX_RC_OFFSET) & BD_TX_RC_MASK);
  65    qemu_log("      Underrun / TCP/IP off-load enable: %d\n",
  66           !!(flags & BD_TX_TOEUN));
  67    qemu_log("      Truncation: %d\n", !!(flags & BD_TX_TR));
  68}
  69
  70static void print_rx_bd_flags(uint16_t flags)
  71{
  72    qemu_log("      Empty: %d\n", !!(flags & BD_RX_EMPTY));
  73    qemu_log("      Receive software ownership: %d\n", !!(flags & BD_RX_RO1));
  74    qemu_log("      Wrap: %d\n", !!(flags & BD_WRAP));
  75    qemu_log("      Interrupt: %d\n", !!(flags & BD_INTERRUPT));
  76    qemu_log("      Last in frame: %d\n", !!(flags & BD_LAST));
  77    qemu_log("      First in frame: %d\n", !!(flags & BD_RX_FIRST));
  78    qemu_log("      Miss: %d\n", !!(flags & BD_RX_MISS));
  79    qemu_log("      Broadcast: %d\n", !!(flags & BD_RX_BROADCAST));
  80    qemu_log("      Multicast: %d\n", !!(flags & BD_RX_MULTICAST));
  81    qemu_log("      Rx frame length violation: %d\n", !!(flags & BD_RX_LG));
  82    qemu_log("      Rx non-octet aligned frame: %d\n", !!(flags & BD_RX_NO));
  83    qemu_log("      Short frame: %d\n", !!(flags & BD_RX_SH));
  84    qemu_log("      Rx CRC Error: %d\n", !!(flags & BD_RX_CR));
  85    qemu_log("      Overrun: %d\n", !!(flags & BD_RX_OV));
  86    qemu_log("      Truncation: %d\n", !!(flags & BD_RX_TR));
  87}
  88
  89
  90static void print_bd(eTSEC_rxtx_bd bd, int mode, uint32_t index)
  91{
  92    qemu_log("eTSEC %s Data Buffer Descriptor (%u)\n",
  93           mode == eTSEC_TRANSMIT ? "Transmit" : "Receive",
  94           index);
  95    qemu_log("   Flags   : 0x%04x\n", bd.flags);
  96    if (mode == eTSEC_TRANSMIT) {
  97        print_tx_bd_flags(bd.flags);
  98    } else {
  99        print_rx_bd_flags(bd.flags);
 100    }
 101    qemu_log("   Length  : 0x%04x\n", bd.length);
 102    qemu_log("   Pointer : 0x%08x\n", bd.bufptr);
 103}
 104
 105#endif  /* DEBUG_BD */
 106
 107static void read_buffer_descriptor(eTSEC         *etsec,
 108                                   hwaddr         addr,
 109                                   eTSEC_rxtx_bd *bd)
 110{
 111    assert(bd != NULL);
 112
 113    RING_DEBUG("READ Buffer Descriptor @ 0x" TARGET_FMT_plx"\n", addr);
 114    cpu_physical_memory_read(addr,
 115                             bd,
 116                             sizeof(eTSEC_rxtx_bd));
 117
 118    if (etsec->regs[DMACTRL].value & DMACTRL_LE) {
 119        bd->flags  = lduw_le_p(&bd->flags);
 120        bd->length = lduw_le_p(&bd->length);
 121        bd->bufptr = ldl_le_p(&bd->bufptr);
 122    } else {
 123        bd->flags  = lduw_be_p(&bd->flags);
 124        bd->length = lduw_be_p(&bd->length);
 125        bd->bufptr = ldl_be_p(&bd->bufptr);
 126    }
 127}
 128
 129static void write_buffer_descriptor(eTSEC         *etsec,
 130                                    hwaddr         addr,
 131                                    eTSEC_rxtx_bd *bd)
 132{
 133    assert(bd != NULL);
 134
 135    if (etsec->regs[DMACTRL].value & DMACTRL_LE) {
 136        stw_le_p(&bd->flags, bd->flags);
 137        stw_le_p(&bd->length, bd->length);
 138        stl_le_p(&bd->bufptr, bd->bufptr);
 139    } else {
 140        stw_be_p(&bd->flags, bd->flags);
 141        stw_be_p(&bd->length, bd->length);
 142        stl_be_p(&bd->bufptr, bd->bufptr);
 143    }
 144
 145    RING_DEBUG("Write Buffer Descriptor @ 0x" TARGET_FMT_plx"\n", addr);
 146    cpu_physical_memory_write(addr,
 147                              bd,
 148                              sizeof(eTSEC_rxtx_bd));
 149}
 150
 151static void ievent_set(eTSEC    *etsec,
 152                       uint32_t  flags)
 153{
 154    etsec->regs[IEVENT].value |= flags;
 155
 156    etsec_update_irq(etsec);
 157}
 158
 159static void tx_padding_and_crc(eTSEC *etsec, uint32_t min_frame_len)
 160{
 161    int add = min_frame_len - etsec->tx_buffer_len;
 162
 163    /* Padding */
 164    if (add > 0) {
 165        RING_DEBUG("pad:%u\n", add);
 166        etsec->tx_buffer = g_realloc(etsec->tx_buffer,
 167                                        etsec->tx_buffer_len + add);
 168
 169        memset(etsec->tx_buffer + etsec->tx_buffer_len, 0x0, add);
 170        etsec->tx_buffer_len += add;
 171    }
 172
 173    /* Never add CRC in QEMU */
 174}
 175
 176static void process_tx_fcb(eTSEC *etsec)
 177{
 178    uint8_t flags = (uint8_t)(*etsec->tx_buffer);
 179    /* L3 header offset from start of frame */
 180    uint8_t l3_header_offset = (uint8_t)*(etsec->tx_buffer + 3);
 181    /* L4 header offset from start of L3 header */
 182    uint8_t l4_header_offset = (uint8_t)*(etsec->tx_buffer + 2);
 183    /* L3 header */
 184    uint8_t *l3_header = etsec->tx_buffer + 8 + l3_header_offset;
 185    /* L4 header */
 186    uint8_t *l4_header = l3_header + l4_header_offset;
 187    int csum = 0;
 188
 189    /* if packet is IP4 and IP checksum is requested */
 190    if (flags & FCB_TX_IP && flags & FCB_TX_CIP) {
 191        csum |= CSUM_IP;
 192    }
 193    /* TODO Check the correct usage of the PHCS field of the FCB in case the NPH
 194     * flag is on */
 195
 196    /* if packet is IP4 and TCP or UDP */
 197    if (flags & FCB_TX_IP && flags & FCB_TX_TUP) {
 198        /* if UDP */
 199        if (flags & FCB_TX_UDP) {
 200            /* if checksum is requested */
 201            if (flags & FCB_TX_CTU) {
 202                /* do UDP checksum */
 203                csum |= CSUM_UDP;
 204            } else {
 205                /* set checksum field to 0 */
 206                l4_header[6] = 0;
 207                l4_header[7] = 0;
 208            }
 209        } else if (flags & FCB_TX_CTU) { /* if TCP and checksum is requested */
 210            /* do TCP checksum */
 211            csum |= CSUM_TCP;
 212        }
 213    }
 214
 215    if (csum) {
 216        net_checksum_calculate(etsec->tx_buffer + 8,
 217                               etsec->tx_buffer_len - 8, csum);
 218    }
 219}
 220
 221static void process_tx_bd(eTSEC         *etsec,
 222                          eTSEC_rxtx_bd *bd)
 223{
 224    uint8_t *tmp_buff = NULL;
 225    hwaddr tbdbth     = (hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32;
 226
 227    if (bd->length == 0) {
 228        /* ERROR */
 229        return;
 230    }
 231
 232    if (etsec->tx_buffer_len == 0) {
 233        /* It's the first BD */
 234        etsec->first_bd = *bd;
 235    }
 236
 237    /* TODO: if TxBD[TOE/UN] skip the Tx Frame Control Block*/
 238
 239    /* Load this Data Buffer */
 240    etsec->tx_buffer = g_realloc(etsec->tx_buffer,
 241                                    etsec->tx_buffer_len + bd->length);
 242    tmp_buff = etsec->tx_buffer + etsec->tx_buffer_len;
 243    cpu_physical_memory_read(bd->bufptr + tbdbth, tmp_buff, bd->length);
 244
 245    /* Update buffer length */
 246    etsec->tx_buffer_len += bd->length;
 247
 248
 249    if (etsec->tx_buffer_len != 0 && (bd->flags & BD_LAST)) {
 250        if (etsec->regs[MACCFG1].value & MACCFG1_TX_EN) {
 251            /* MAC Transmit enabled */
 252
 253            /* Process offload Tx FCB */
 254            if (etsec->first_bd.flags & BD_TX_TOEUN) {
 255                process_tx_fcb(etsec);
 256            }
 257
 258            if (etsec->first_bd.flags & BD_TX_PADCRC
 259                || etsec->regs[MACCFG2].value & MACCFG2_PADCRC) {
 260
 261                /* Padding and CRC (Padding implies CRC) */
 262                tx_padding_and_crc(etsec, 60);
 263
 264            } else if (etsec->first_bd.flags & BD_TX_TC
 265                       || etsec->regs[MACCFG2].value & MACCFG2_CRC_EN) {
 266
 267                /* Only CRC */
 268                /* Never add CRC in QEMU */
 269            }
 270
 271#if defined(HEX_DUMP)
 272            qemu_log("eTSEC Send packet size:%d\n", etsec->tx_buffer_len);
 273            qemu_hexdump(stderr, "", etsec->tx_buffer, etsec->tx_buffer_len);
 274#endif  /* ETSEC_RING_DEBUG */
 275
 276            if (etsec->first_bd.flags & BD_TX_TOEUN) {
 277                qemu_send_packet(qemu_get_queue(etsec->nic),
 278                        etsec->tx_buffer + 8,
 279                        etsec->tx_buffer_len - 8);
 280            } else {
 281                qemu_send_packet(qemu_get_queue(etsec->nic),
 282                        etsec->tx_buffer,
 283                        etsec->tx_buffer_len);
 284            }
 285
 286        }
 287
 288        etsec->tx_buffer_len = 0;
 289
 290        if (bd->flags & BD_INTERRUPT) {
 291            ievent_set(etsec, IEVENT_TXF);
 292        }
 293    } else {
 294        if (bd->flags & BD_INTERRUPT) {
 295            ievent_set(etsec, IEVENT_TXB);
 296        }
 297    }
 298
 299    /* Update DB flags */
 300
 301    /* Clear Ready */
 302    bd->flags &= ~BD_TX_READY;
 303
 304    /* Clear Defer */
 305    bd->flags &= ~BD_TX_PREDEF;
 306
 307    /* Clear Late Collision */
 308    bd->flags &= ~BD_TX_HFELC;
 309
 310    /* Clear Retransmission Limit */
 311    bd->flags &= ~BD_TX_CFRL;
 312
 313    /* Clear Retry Count */
 314    bd->flags &= ~(BD_TX_RC_MASK << BD_TX_RC_OFFSET);
 315
 316    /* Clear Underrun */
 317    bd->flags &= ~BD_TX_TOEUN;
 318
 319    /* Clear Truncation */
 320    bd->flags &= ~BD_TX_TR;
 321}
 322
 323void etsec_walk_tx_ring(eTSEC *etsec, int ring_nbr)
 324{
 325    hwaddr        ring_base = 0;
 326    hwaddr        bd_addr   = 0;
 327    eTSEC_rxtx_bd bd;
 328    uint16_t      bd_flags;
 329
 330    if (!(etsec->regs[MACCFG1].value & MACCFG1_TX_EN)) {
 331        RING_DEBUG("%s: MAC Transmit not enabled\n", __func__);
 332        return;
 333    }
 334
 335    ring_base = (hwaddr)(etsec->regs[TBASEH].value & 0xF) << 32;
 336    ring_base += etsec->regs[TBASE0 + ring_nbr].value & ~0x7;
 337    bd_addr    = etsec->regs[TBPTR0 + ring_nbr].value & ~0x7;
 338
 339    do {
 340        read_buffer_descriptor(etsec, bd_addr, &bd);
 341
 342#ifdef DEBUG_BD
 343        print_bd(bd,
 344                 eTSEC_TRANSMIT,
 345                 (bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd));
 346
 347#endif  /* DEBUG_BD */
 348
 349        /* Save flags before BD update */
 350        bd_flags = bd.flags;
 351
 352        if (!(bd_flags & BD_TX_READY)) {
 353            break;
 354        }
 355
 356        process_tx_bd(etsec, &bd);
 357        /* Write back BD after update */
 358        write_buffer_descriptor(etsec, bd_addr, &bd);
 359
 360        /* Wrap or next BD */
 361        if (bd_flags & BD_WRAP) {
 362            bd_addr = ring_base;
 363        } else {
 364            bd_addr += sizeof(eTSEC_rxtx_bd);
 365        }
 366    } while (TRUE);
 367
 368    /* Save the Buffer Descriptor Pointers to last bd that was not
 369     * succesfully closed */
 370    etsec->regs[TBPTR0 + ring_nbr].value = bd_addr;
 371
 372    /* Set transmit halt THLTx */
 373    etsec->regs[TSTAT].value |= 1 << (31 - ring_nbr);
 374}
 375
 376static void fill_rx_bd(eTSEC          *etsec,
 377                       eTSEC_rxtx_bd  *bd,
 378                       const uint8_t **buf,
 379                       size_t         *size)
 380{
 381    uint16_t to_write;
 382    hwaddr   bufptr = bd->bufptr +
 383        ((hwaddr)(etsec->regs[TBDBPH].value & 0xF) << 32);
 384    uint8_t  padd[etsec->rx_padding];
 385    uint8_t  rem;
 386
 387    RING_DEBUG("eTSEC fill Rx buffer @ 0x%016" HWADDR_PRIx
 388               " size:%zu(padding + crc:%u) + fcb:%u\n",
 389               bufptr, *size, etsec->rx_padding, etsec->rx_fcb_size);
 390
 391    bd->length = 0;
 392
 393    /* This operation will only write FCB */
 394    if (etsec->rx_fcb_size != 0) {
 395
 396        cpu_physical_memory_write(bufptr, etsec->rx_fcb, etsec->rx_fcb_size);
 397
 398        bufptr             += etsec->rx_fcb_size;
 399        bd->length         += etsec->rx_fcb_size;
 400        etsec->rx_fcb_size  = 0;
 401
 402    }
 403
 404    /* We remove padding from the computation of to_write because it is not
 405     * allocated in the buffer.
 406     */
 407    to_write = MIN(*size - etsec->rx_padding,
 408                   etsec->regs[MRBLR].value - etsec->rx_fcb_size);
 409
 410    /* This operation can only write packet data and no padding */
 411    if (to_write > 0) {
 412        cpu_physical_memory_write(bufptr, *buf, to_write);
 413
 414        *buf   += to_write;
 415        bufptr += to_write;
 416        *size  -= to_write;
 417
 418        bd->flags  &= ~BD_RX_EMPTY;
 419        bd->length += to_write;
 420    }
 421
 422    if (*size == etsec->rx_padding) {
 423        /* The remaining bytes are only for padding which is not actually
 424         * allocated in the data buffer.
 425         */
 426
 427        rem = MIN(etsec->regs[MRBLR].value - bd->length, etsec->rx_padding);
 428
 429        if (rem > 0) {
 430            memset(padd, 0x0, sizeof(padd));
 431            etsec->rx_padding -= rem;
 432            *size             -= rem;
 433            bd->length        += rem;
 434            cpu_physical_memory_write(bufptr, padd, rem);
 435        }
 436    }
 437}
 438
 439static void rx_init_frame(eTSEC *etsec, const uint8_t *buf, size_t size)
 440{
 441    uint32_t fcb_size = 0;
 442    uint8_t  prsdep   = (etsec->regs[RCTRL].value >> RCTRL_PRSDEP_OFFSET)
 443        & RCTRL_PRSDEP_MASK;
 444
 445    if (prsdep != 0) {
 446        /* Prepend FCB (FCB size + RCTRL[PAL]) */
 447        fcb_size = 8 + ((etsec->regs[RCTRL].value >> 16) & 0x1F);
 448
 449        etsec->rx_fcb_size = fcb_size;
 450
 451        /* TODO: fill_FCB(etsec); */
 452        memset(etsec->rx_fcb, 0x0, sizeof(etsec->rx_fcb));
 453
 454    } else {
 455        etsec->rx_fcb_size = 0;
 456    }
 457
 458    g_free(etsec->rx_buffer);
 459
 460    /* Do not copy the frame for now */
 461    etsec->rx_buffer     = (uint8_t *)buf;
 462    etsec->rx_buffer_len = size;
 463
 464    /* CRC padding (We don't have to compute the CRC) */
 465    etsec->rx_padding = 4;
 466
 467    /*
 468     * Ensure that payload length + CRC length is at least 802.3
 469     * minimum MTU size bytes long (64)
 470     */
 471    if (etsec->rx_buffer_len < 60) {
 472        etsec->rx_padding += 60 - etsec->rx_buffer_len;
 473    }
 474
 475    etsec->rx_first_in_frame = 1;
 476    etsec->rx_remaining_data = etsec->rx_buffer_len;
 477    RING_DEBUG("%s: rx_buffer_len:%u rx_padding+crc:%u\n", __func__,
 478               etsec->rx_buffer_len, etsec->rx_padding);
 479}
 480
 481ssize_t etsec_rx_ring_write(eTSEC *etsec, const uint8_t *buf, size_t size)
 482{
 483    int ring_nbr = 0;           /* Always use ring0 (no filer) */
 484
 485    if (etsec->rx_buffer_len != 0) {
 486        RING_DEBUG("%s: We can't receive now,"
 487                   " a buffer is already in the pipe\n", __func__);
 488        return 0;
 489    }
 490
 491    if (etsec->regs[RSTAT].value & 1 << (23 - ring_nbr)) {
 492        RING_DEBUG("%s: The ring is halted\n", __func__);
 493        return -1;
 494    }
 495
 496    if (etsec->regs[DMACTRL].value & DMACTRL_GRS) {
 497        RING_DEBUG("%s: Graceful receive stop\n", __func__);
 498        return -1;
 499    }
 500
 501    if (!(etsec->regs[MACCFG1].value & MACCFG1_RX_EN)) {
 502        RING_DEBUG("%s: MAC Receive not enabled\n", __func__);
 503        return -1;
 504    }
 505
 506    if (!(etsec->regs[RCTRL].value & RCTRL_RSF) && (size < 60)) {
 507        /* CRC is not in the packet yet, so short frame is below 60 bytes */
 508        RING_DEBUG("%s: Drop short frame\n", __func__);
 509        return -1;
 510    }
 511
 512    rx_init_frame(etsec, buf, size);
 513
 514    etsec_walk_rx_ring(etsec, ring_nbr);
 515
 516    return size;
 517}
 518
 519void etsec_walk_rx_ring(eTSEC *etsec, int ring_nbr)
 520{
 521    hwaddr         ring_base     = 0;
 522    hwaddr         bd_addr       = 0;
 523    hwaddr         start_bd_addr = 0;
 524    eTSEC_rxtx_bd  bd;
 525    uint16_t       bd_flags;
 526    size_t         remaining_data;
 527    const uint8_t *buf;
 528    uint8_t       *tmp_buf;
 529    size_t         size;
 530
 531    if (etsec->rx_buffer_len == 0) {
 532        /* No frame to send */
 533        RING_DEBUG("No frame to send\n");
 534        return;
 535    }
 536
 537    remaining_data = etsec->rx_remaining_data + etsec->rx_padding;
 538    buf            = etsec->rx_buffer
 539        + (etsec->rx_buffer_len - etsec->rx_remaining_data);
 540    size           = etsec->rx_buffer_len + etsec->rx_padding;
 541
 542    ring_base = (hwaddr)(etsec->regs[RBASEH].value & 0xF) << 32;
 543    ring_base += etsec->regs[RBASE0 + ring_nbr].value & ~0x7;
 544    start_bd_addr  = bd_addr = etsec->regs[RBPTR0 + ring_nbr].value & ~0x7;
 545
 546    do {
 547        read_buffer_descriptor(etsec, bd_addr, &bd);
 548
 549#ifdef DEBUG_BD
 550        print_bd(bd,
 551                 eTSEC_RECEIVE,
 552                 (bd_addr - ring_base) / sizeof(eTSEC_rxtx_bd));
 553
 554#endif  /* DEBUG_BD */
 555
 556        /* Save flags before BD update */
 557        bd_flags = bd.flags;
 558
 559        if (bd_flags & BD_RX_EMPTY) {
 560            fill_rx_bd(etsec, &bd, &buf, &remaining_data);
 561
 562            if (etsec->rx_first_in_frame) {
 563                bd.flags |= BD_RX_FIRST;
 564                etsec->rx_first_in_frame = 0;
 565                etsec->rx_first_bd = bd;
 566            }
 567
 568            /* Last in frame */
 569            if (remaining_data == 0) {
 570
 571                /* Clear flags */
 572
 573                bd.flags &= ~0x7ff;
 574
 575                bd.flags |= BD_LAST;
 576
 577                /* NOTE: non-octet aligned frame is impossible in qemu */
 578
 579                if (size >= etsec->regs[MAXFRM].value) {
 580                    /* frame length violation */
 581                    qemu_log("%s frame length violation: size:%zu MAXFRM:%d\n",
 582                           __func__, size, etsec->regs[MAXFRM].value);
 583
 584                    bd.flags |= BD_RX_LG;
 585                }
 586
 587                if (size  < 64) {
 588                    /* Short frame */
 589                    bd.flags |= BD_RX_SH;
 590                }
 591
 592                /* TODO: Broadcast and Multicast */
 593
 594                if (bd.flags & BD_INTERRUPT) {
 595                    /* Set RXFx */
 596                    etsec->regs[RSTAT].value |= 1 << (7 - ring_nbr);
 597
 598                    /* Set IEVENT */
 599                    ievent_set(etsec, IEVENT_RXF);
 600                }
 601
 602            } else {
 603                if (bd.flags & BD_INTERRUPT) {
 604                    /* Set IEVENT */
 605                    ievent_set(etsec, IEVENT_RXB);
 606                }
 607            }
 608
 609            /* Write back BD after update */
 610            write_buffer_descriptor(etsec, bd_addr, &bd);
 611        }
 612
 613        /* Wrap or next BD */
 614        if (bd_flags & BD_WRAP) {
 615            bd_addr = ring_base;
 616        } else {
 617            bd_addr += sizeof(eTSEC_rxtx_bd);
 618        }
 619    } while (remaining_data != 0
 620             && (bd_flags & BD_RX_EMPTY)
 621             && bd_addr != start_bd_addr);
 622
 623    /* Reset ring ptr */
 624    etsec->regs[RBPTR0 + ring_nbr].value = bd_addr;
 625
 626    /* The frame is too large to fit in the Rx ring */
 627    if (remaining_data > 0) {
 628
 629        /* Set RSTAT[QHLTx] */
 630        etsec->regs[RSTAT].value |= 1 << (23 - ring_nbr);
 631
 632        /* Save remaining data to send the end of the frame when the ring will
 633         * be restarted
 634         */
 635        etsec->rx_remaining_data = remaining_data;
 636
 637        /* Copy the frame */
 638        tmp_buf = g_malloc(size);
 639        memcpy(tmp_buf, etsec->rx_buffer, size);
 640        etsec->rx_buffer = tmp_buf;
 641
 642        RING_DEBUG("no empty RxBD available any more\n");
 643    } else {
 644        etsec->rx_buffer_len = 0;
 645        etsec->rx_buffer     = NULL;
 646        if (etsec->need_flush) {
 647            qemu_flush_queued_packets(qemu_get_queue(etsec->nic));
 648        }
 649    }
 650
 651    RING_DEBUG("eTSEC End of ring_write: remaining_data:%zu\n", remaining_data);
 652}
 653