@@ -668,6 +668,7 @@ struct rtl8169_counters {
};
enum rtl_flag {
+ RTL_FLAG_TASK_ENABLED,
RTL_FLAG_TASK_SLOW_PENDING,
RTL_FLAG_TASK_RESET_PENDING,
RTL_FLAG_TASK_PHY_PENDING,
@@ -679,7 +680,6 @@ struct rtl8169_private {
struct pci_dev *pci_dev;
struct net_device *dev;
struct napi_struct napi;
- spinlock_t lock;
u32 msg_enable;
u16 txd_version;
u16 mac_version;
@@ -725,7 +725,6 @@ struct rtl8169_private {
DECLARE_BITMAP(flags, RTL_FLAG_MAX);
struct mutex mutex;
struct work_struct work;
- bool enabled;
} wk;
unsigned features;
@@ -1723,9 +1722,7 @@ static void rtl8169_get_regs(struct net_device *dev, struct ethtool_regs *regs,
regs->len = R8169_REGS_SIZE;
rtl_lock_work(tp);
- spin_lock_bh(&tp->lock);
memcpy_fromio(p, tp->mmio_addr, regs->len);
- spin_unlock_bh(&tp->lock);
rtl_unlock_work(tp);
}
@@ -3273,17 +3270,8 @@ out_mod_timer:
static void rtl_schedule_task(struct rtl8169_private *tp, enum rtl_flag flag)
{
- spin_lock(&tp->lock);
if (!test_and_set_bit(flag, tp->wk.flags))
schedule_work(&tp->wk.work);
- spin_unlock(&tp->lock);
-}
-
-static void rtl_schedule_task_bh(struct rtl8169_private *tp, enum rtl_flag flag)
-{
- local_bh_disable();
- rtl_schedule_task(tp, flag);
- local_bh_enable();
}
static void rtl8169_phy_timer(unsigned long __opaque)
@@ -3291,7 +3279,7 @@ static void rtl8169_phy_timer(unsigned long __opaque)
struct net_device *dev = (struct net_device *)__opaque;
struct rtl8169_private *tp = netdev_priv(dev);
- rtl_schedule_task_bh(tp, RTL_FLAG_TASK_PHY_PENDING);
+ rtl_schedule_task(tp, RTL_FLAG_TASK_PHY_PENDING);
}
#ifdef CONFIG_NET_POLL_CONTROLLER
@@ -4160,7 +4148,6 @@ rtl8169_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
tp->do_ioctl = rtl_xmii_ioctl;
}
- spin_lock_init(&tp->lock);
mutex_init(&tp->wk.mutex);
/* Get MAC address */
@@ -4361,7 +4348,7 @@ static int rtl8169_open(struct net_device *dev)
rtl_lock_work(tp);
- tp->wk.enabled = true;
+ set_bit(RTL_FLAG_TASK_ENABLED, tp->wk.flags);
napi_enable(&tp->napi);
@@ -5559,7 +5546,18 @@ static netdev_tx_t rtl8169_start_xmit(struct sk_buff *skb,
mmiowb();
if (TX_BUFFS_AVAIL(tp) < MAX_SKB_FRAGS) {
+ /* Avoid wrongly optimistic queue wake-up: rtl_tx thread must
+ * not miss a ring update when it notices a stopped queue.
+ */
+ smp_wmb();
netif_stop_queue(dev);
+ /* Sync with rtl_tx:
+ * - publish queue status and cur_tx ring index (write barrier)
+ * - refresh dirty_tx ring index (read barrier).
+ * May the current thread have a pessimistic view of the ring
+ * status and forget to wake up queue, a racing rtl_tx thread
+ * can't.
+ */
smp_mb();
if (TX_BUFFS_AVAIL(tp) >= MAX_SKB_FRAGS)
netif_wake_queue(dev);
@@ -5624,7 +5622,7 @@ static void rtl8169_pcierr_interrupt(struct net_device *dev)
rtl8169_hw_reset(tp);
- rtl_schedule_task_bh(tp, RTL_FLAG_TASK_RESET_PENDING);
+ rtl_schedule_task(tp, RTL_FLAG_TASK_RESET_PENDING);
}
static void rtl_tx(struct net_device *dev, struct rtl8169_private *tp)
@@ -5659,6 +5657,13 @@ static void rtl_tx(struct net_device *dev, struct rtl8169_private *tp)
if (tp->dirty_tx != dirty_tx) {
tp->dirty_tx = dirty_tx;
+ /* Sync with rtl8169_start_xmit:
+ * - publish dirty_tx ring index (write barrier)
+ * - refresh cur_tx ring index and queue status (read barrier)
+ * May the current thread miss the stopped queue condition,
+ * a racing xmit thread can only have a right view of the
+ * ring status.
+ */
smp_mb();
if (netif_queue_stopped(dev) &&
(TX_BUFFS_AVAIL(tp) >= MAX_SKB_FRAGS)) {
@@ -5834,7 +5839,8 @@ static void rtl_slow_event_work(struct rtl8169_private *tp)
/* Work around for rx fifo overflow */
case RTL_GIGA_MAC_VER_11:
netif_stop_queue(dev);
- rtl_schedule_task_bh(tp, RTL_FLAG_TASK_RESET_PENDING);
+ /* XXX - Hack alert. See rtl_task(). */
+ set_bit(RTL_FLAG_TASK_RESET_PENDING, tp->wk.flags);
default:
break;
}
@@ -5859,6 +5865,7 @@ static void rtl_task(struct work_struct *work)
int bitnr;
void (*action)(struct rtl8169_private *);
} rtl_work[] = {
+ /* XXX - keep rtl_slow_event_work() as first element. */
{ RTL_FLAG_TASK_SLOW_PENDING, rtl_slow_event_work },
{ RTL_FLAG_TASK_RESET_PENDING, rtl_reset_work },
{ RTL_FLAG_TASK_PHY_PENDING, rtl_phy_work }
@@ -5870,16 +5877,14 @@ static void rtl_task(struct work_struct *work)
rtl_lock_work(tp);
- if (!netif_running(dev) || !tp->wk.enabled)
+ if (!netif_running(dev) ||
+ !test_bit(RTL_FLAG_TASK_ENABLED, tp->wk.flags))
goto out_unlock;
for (i = 0; i < ARRAY_SIZE(rtl_work); i++) {
bool pending;
- spin_lock_bh(&tp->lock);
pending = test_and_clear_bit(rtl_work[i].bitnr, tp->wk.flags);
- spin_unlock_bh(&tp->lock);
-
if (pending)
rtl_work[i].action(tp);
}
@@ -5971,7 +5976,7 @@ static int rtl8169_close(struct net_device *dev)
rtl8169_update_counters(dev);
rtl_lock_work(tp);
- tp->wk.enabled = false;
+ clear_bit(RTL_FLAG_TASK_ENABLED, tp->wk.flags);
rtl8169_down(dev);
rtl_unlock_work(tp);
@@ -6022,8 +6027,6 @@ static void rtl_set_rx_mode(struct net_device *dev)
}
}
- spin_lock_bh(&tp->lock);
-
tmp = (RTL_R32(RxConfig) & ~RX_CONFIG_ACCEPT_MASK) | rx_mode;
if (tp->mac_version > RTL_GIGA_MAC_VER_06) {
@@ -6037,8 +6040,6 @@ static void rtl_set_rx_mode(struct net_device *dev)
RTL_W32(MAR0 + 0, mc_filter[0]);
RTL_W32(RxConfig, tmp);
-
- spin_unlock_bh(&tp->lock);
}
/**
@@ -6070,7 +6071,7 @@ static void rtl8169_net_suspend(struct net_device *dev)
rtl_lock_work(tp);
napi_disable(&tp->napi);
- tp->wk.enabled = false;
+ clear_bit(RTL_FLAG_TASK_ENABLED, tp->wk.flags);
rtl_unlock_work(tp);
rtl_pll_power_down(tp);
@@ -6096,9 +6097,9 @@ static void __rtl8169_resume(struct net_device *dev)
rtl_pll_power_up(tp);
- tp->wk.enabled = true;
+ set_bit(RTL_FLAG_TASK_ENABLED, tp->wk.flags);
- rtl_schedule_task_bh(tp, RTL_FLAG_TASK_RESET_PENDING);
+ rtl_schedule_task(tp, RTL_FLAG_TASK_RESET_PENDING);
}
static int rtl8169_resume(struct device *device)