struct page *page,
u32 page_offset,
u32 len,
- bool hidma)
+#ifdef __NO_BOOL__
+ u8 hidma
+#else
+ bool hidma
+#endif
+ )
{
dma_addr_t dmaaddr;
struct jme_ring *txring = &(jme->txring[0]);
struct txdesc *txdesc = txring->desc, *ctxdesc;
struct jme_buffer_info *txbi = txring->bufinf, *ctxbi;
+#ifdef __NO_BOOL__
+ u8 hidma = !!(jme->dev->features & NETIF_F_HIGHDMA);
+#else
bool hidma = jme->dev->features & NETIF_F_HIGHDMA;
+#endif
int i, nr_frags = skb_shinfo(skb)->nr_frags;
int mask = jme->tx_ring_mask;
const struct skb_frag_struct *frag;
((new_mtu) < IPV6_MIN_MTU))
return -EINVAL;
- if (new_mtu > 4000) {
- jme->reg_rxcs &= ~RXCS_FIFOTHNP;
- jme->reg_rxcs |= RXCS_FIFOTHNP_64QW;
- jme_restart_rx_engine(jme);
- } else {
- jme->reg_rxcs &= ~RXCS_FIFOTHNP;
- jme->reg_rxcs |= RXCS_FIFOTHNP_128QW;
- jme_restart_rx_engine(jme);
- }
#ifndef __USE_NDO_FIX_FEATURES__
if (new_mtu > 1900) {
#ifdef __USE_NDO_FIX_FEATURES__
netdev_update_features(netdev);
#endif
+
+ jme_restart_rx_engine(jme);
jme_reset_link(jme);
return 0;
*/
netdev = alloc_etherdev(sizeof(*jme));
if (!netdev) {
- pr_err("Cannot allocate netdev structure\n");
rc = -ENOMEM;
goto err_out_release_regions;
}