forked from Minki/linux
ide: use per-device request queue locks (v2)
* Move hack for flush requests from choose_drive() to do_ide_request(). * Add ide_plug_device() helper and convert core IDE code from using per-hwgroup lock as a request lock to use the ->queue_lock instead. * Remove no longer needed: - choose_drive() function - WAKEUP() macro - 'sleeping' flag from ide_hwif_t - 'service_{start,time}' fields from ide_drive_t This patch results in much simpler and more maintainable code (besides being a scalability improvement). v2: * Fixes/improvements based on review from Elias: - take as many requests off the queue as possible - remove now redundant BUG_ON() Cc: Elias Oltmanns <eo@nebensachen.de> Signed-off-by: Bartlomiej Zolnierkiewicz <bzolnier@gmail.com>
This commit is contained in:
parent
631de3708d
commit
201bffa464
@ -667,85 +667,10 @@ void ide_stall_queue (ide_drive_t *drive, unsigned long timeout)
|
|||||||
drive->sleep = timeout + jiffies;
|
drive->sleep = timeout + jiffies;
|
||||||
drive->dev_flags |= IDE_DFLAG_SLEEPING;
|
drive->dev_flags |= IDE_DFLAG_SLEEPING;
|
||||||
}
|
}
|
||||||
|
|
||||||
EXPORT_SYMBOL(ide_stall_queue);
|
EXPORT_SYMBOL(ide_stall_queue);
|
||||||
|
|
||||||
#define WAKEUP(drive) ((drive)->service_start + 2 * (drive)->service_time)
|
|
||||||
|
|
||||||
/**
|
|
||||||
* choose_drive - select a drive to service
|
|
||||||
* @hwgroup: hardware group to select on
|
|
||||||
*
|
|
||||||
* choose_drive() selects the next drive which will be serviced.
|
|
||||||
* This is necessary because the IDE layer can't issue commands
|
|
||||||
* to both drives on the same cable, unlike SCSI.
|
|
||||||
*/
|
|
||||||
|
|
||||||
static inline ide_drive_t *choose_drive (ide_hwgroup_t *hwgroup)
|
|
||||||
{
|
|
||||||
ide_drive_t *drive, *best;
|
|
||||||
|
|
||||||
repeat:
|
|
||||||
best = NULL;
|
|
||||||
drive = hwgroup->drive;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* drive is doing pre-flush, ordered write, post-flush sequence. even
|
|
||||||
* though that is 3 requests, it must be seen as a single transaction.
|
|
||||||
* we must not preempt this drive until that is complete
|
|
||||||
*/
|
|
||||||
if (blk_queue_flushing(drive->queue)) {
|
|
||||||
/*
|
|
||||||
* small race where queue could get replugged during
|
|
||||||
* the 3-request flush cycle, just yank the plug since
|
|
||||||
* we want it to finish asap
|
|
||||||
*/
|
|
||||||
blk_remove_plug(drive->queue);
|
|
||||||
return drive;
|
|
||||||
}
|
|
||||||
|
|
||||||
do {
|
|
||||||
u8 dev_s = !!(drive->dev_flags & IDE_DFLAG_SLEEPING);
|
|
||||||
u8 best_s = (best && !!(best->dev_flags & IDE_DFLAG_SLEEPING));
|
|
||||||
|
|
||||||
if ((dev_s == 0 || time_after_eq(jiffies, drive->sleep)) &&
|
|
||||||
!elv_queue_empty(drive->queue)) {
|
|
||||||
if (best == NULL ||
|
|
||||||
(dev_s && (best_s == 0 || time_before(drive->sleep, best->sleep))) ||
|
|
||||||
(best_s == 0 && time_before(WAKEUP(drive), WAKEUP(best)))) {
|
|
||||||
if (!blk_queue_plugged(drive->queue))
|
|
||||||
best = drive;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} while ((drive = drive->next) != hwgroup->drive);
|
|
||||||
|
|
||||||
if (best && (best->dev_flags & IDE_DFLAG_NICE1) &&
|
|
||||||
(best->dev_flags & IDE_DFLAG_SLEEPING) == 0 &&
|
|
||||||
best != hwgroup->drive && best->service_time > WAIT_MIN_SLEEP) {
|
|
||||||
long t = (signed long)(WAKEUP(best) - jiffies);
|
|
||||||
if (t >= WAIT_MIN_SLEEP) {
|
|
||||||
/*
|
|
||||||
* We *may* have some time to spare, but first let's see if
|
|
||||||
* someone can potentially benefit from our nice mood today..
|
|
||||||
*/
|
|
||||||
drive = best->next;
|
|
||||||
do {
|
|
||||||
if ((drive->dev_flags & IDE_DFLAG_SLEEPING) == 0
|
|
||||||
&& time_before(jiffies - best->service_time, WAKEUP(drive))
|
|
||||||
&& time_before(WAKEUP(drive), jiffies + t))
|
|
||||||
{
|
|
||||||
ide_stall_queue(best, min_t(long, t, 10 * WAIT_MIN_SLEEP));
|
|
||||||
goto repeat;
|
|
||||||
}
|
|
||||||
} while ((drive = drive->next) != best);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return best;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Issue a new request to a drive from hwgroup
|
* Issue a new request to a drive from hwgroup
|
||||||
* Caller must have already done spin_lock_irqsave(&hwgroup->lock, ..);
|
|
||||||
*
|
*
|
||||||
* A hwgroup is a serialized group of IDE interfaces. Usually there is
|
* A hwgroup is a serialized group of IDE interfaces. Usually there is
|
||||||
* exactly one hwif (interface) per hwgroup, but buggy controllers (eg. CMD640)
|
* exactly one hwif (interface) per hwgroup, but buggy controllers (eg. CMD640)
|
||||||
@ -757,8 +682,7 @@ repeat:
|
|||||||
* possibly along with many other devices. This is especially common in
|
* possibly along with many other devices. This is especially common in
|
||||||
* PCI-based systems with off-board IDE controller cards.
|
* PCI-based systems with off-board IDE controller cards.
|
||||||
*
|
*
|
||||||
* The IDE driver uses a per-hwgroup spinlock to protect
|
* The IDE driver uses a per-hwgroup lock to protect the hwgroup->busy flag.
|
||||||
* access to the request queues, and to protect the hwgroup->busy flag.
|
|
||||||
*
|
*
|
||||||
* The first thread into the driver for a particular hwgroup sets the
|
* The first thread into the driver for a particular hwgroup sets the
|
||||||
* hwgroup->busy flag to indicate that this hwgroup is now active,
|
* hwgroup->busy flag to indicate that this hwgroup is now active,
|
||||||
@ -780,62 +704,39 @@ repeat:
|
|||||||
*/
|
*/
|
||||||
void do_ide_request(struct request_queue *q)
|
void do_ide_request(struct request_queue *q)
|
||||||
{
|
{
|
||||||
ide_drive_t *orig_drive = q->queuedata;
|
ide_drive_t *drive = q->queuedata;
|
||||||
ide_hwgroup_t *hwgroup = orig_drive->hwif->hwgroup;
|
ide_hwif_t *hwif = drive->hwif;
|
||||||
ide_drive_t *drive;
|
ide_hwgroup_t *hwgroup = hwif->hwgroup;
|
||||||
ide_hwif_t *hwif;
|
|
||||||
struct request *rq;
|
struct request *rq;
|
||||||
ide_startstop_t startstop;
|
ide_startstop_t startstop;
|
||||||
|
|
||||||
/* caller must own hwgroup->lock */
|
/*
|
||||||
BUG_ON(!irqs_disabled());
|
* drive is doing pre-flush, ordered write, post-flush sequence. even
|
||||||
|
* though that is 3 requests, it must be seen as a single transaction.
|
||||||
while (!ide_lock_hwgroup(hwgroup)) {
|
* we must not preempt this drive until that is complete
|
||||||
drive = choose_drive(hwgroup);
|
*/
|
||||||
if (drive == NULL) {
|
if (blk_queue_flushing(q))
|
||||||
int sleeping = 0;
|
|
||||||
unsigned long sleep = 0; /* shut up, gcc */
|
|
||||||
hwgroup->rq = NULL;
|
|
||||||
drive = hwgroup->drive;
|
|
||||||
do {
|
|
||||||
if ((drive->dev_flags & IDE_DFLAG_SLEEPING) &&
|
|
||||||
(sleeping == 0 ||
|
|
||||||
time_before(drive->sleep, sleep))) {
|
|
||||||
sleeping = 1;
|
|
||||||
sleep = drive->sleep;
|
|
||||||
}
|
|
||||||
} while ((drive = drive->next) != hwgroup->drive);
|
|
||||||
if (sleeping) {
|
|
||||||
/*
|
/*
|
||||||
* Take a short snooze, and then wake up this hwgroup again.
|
* small race where queue could get replugged during
|
||||||
* This gives other hwgroups on the same a chance to
|
* the 3-request flush cycle, just yank the plug since
|
||||||
* play fairly with us, just in case there are big differences
|
* we want it to finish asap
|
||||||
* in relative throughputs.. don't want to hog the cpu too much.
|
|
||||||
*/
|
*/
|
||||||
if (time_before(sleep, jiffies + WAIT_MIN_SLEEP))
|
blk_remove_plug(q);
|
||||||
sleep = jiffies + WAIT_MIN_SLEEP;
|
|
||||||
#if 1
|
spin_unlock_irq(q->queue_lock);
|
||||||
if (timer_pending(&hwgroup->timer))
|
spin_lock_irq(&hwgroup->lock);
|
||||||
printk(KERN_CRIT "ide_set_handler: timer already active\n");
|
|
||||||
#endif
|
if (!ide_lock_hwgroup(hwgroup)) {
|
||||||
/* so that ide_timer_expiry knows what to do */
|
repeat:
|
||||||
hwgroup->sleeping = 1;
|
hwgroup->rq = NULL;
|
||||||
hwgroup->req_gen_timer = hwgroup->req_gen;
|
|
||||||
mod_timer(&hwgroup->timer, sleep);
|
if (drive->dev_flags & IDE_DFLAG_SLEEPING) {
|
||||||
/* we purposely leave hwgroup locked
|
if (time_before(drive->sleep, jiffies)) {
|
||||||
* while sleeping */
|
|
||||||
} else
|
|
||||||
ide_unlock_hwgroup(hwgroup);
|
ide_unlock_hwgroup(hwgroup);
|
||||||
|
goto plug_device;
|
||||||
/* no more work for this hwgroup (for now) */
|
}
|
||||||
goto plug_device;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (drive != orig_drive)
|
|
||||||
goto plug_device;
|
|
||||||
|
|
||||||
hwif = drive->hwif;
|
|
||||||
|
|
||||||
if (hwif != hwgroup->hwif) {
|
if (hwif != hwgroup->hwif) {
|
||||||
/*
|
/*
|
||||||
* set nIEN for previous hwif, drives in the
|
* set nIEN for previous hwif, drives in the
|
||||||
@ -847,16 +748,20 @@ void do_ide_request(struct request_queue *q)
|
|||||||
hwgroup->hwif = hwif;
|
hwgroup->hwif = hwif;
|
||||||
hwgroup->drive = drive;
|
hwgroup->drive = drive;
|
||||||
drive->dev_flags &= ~(IDE_DFLAG_SLEEPING | IDE_DFLAG_PARKED);
|
drive->dev_flags &= ~(IDE_DFLAG_SLEEPING | IDE_DFLAG_PARKED);
|
||||||
drive->service_start = jiffies;
|
|
||||||
|
|
||||||
|
spin_unlock_irq(&hwgroup->lock);
|
||||||
|
spin_lock_irq(q->queue_lock);
|
||||||
/*
|
/*
|
||||||
* we know that the queue isn't empty, but this can happen
|
* we know that the queue isn't empty, but this can happen
|
||||||
* if the q->prep_rq_fn() decides to kill a request
|
* if the q->prep_rq_fn() decides to kill a request
|
||||||
*/
|
*/
|
||||||
rq = elv_next_request(drive->queue);
|
rq = elv_next_request(drive->queue);
|
||||||
|
spin_unlock_irq(q->queue_lock);
|
||||||
|
spin_lock_irq(&hwgroup->lock);
|
||||||
|
|
||||||
if (!rq) {
|
if (!rq) {
|
||||||
ide_unlock_hwgroup(hwgroup);
|
ide_unlock_hwgroup(hwgroup);
|
||||||
break;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -886,17 +791,21 @@ void do_ide_request(struct request_queue *q)
|
|||||||
startstop = start_request(drive, rq);
|
startstop = start_request(drive, rq);
|
||||||
spin_lock_irq(&hwgroup->lock);
|
spin_lock_irq(&hwgroup->lock);
|
||||||
|
|
||||||
if (startstop == ide_stopped) {
|
if (startstop == ide_stopped)
|
||||||
ide_unlock_hwgroup(hwgroup);
|
goto repeat;
|
||||||
if (!elv_queue_empty(orig_drive->queue))
|
} else
|
||||||
blk_plug_device(orig_drive->queue);
|
goto plug_device;
|
||||||
}
|
out:
|
||||||
}
|
spin_unlock_irq(&hwgroup->lock);
|
||||||
|
spin_lock_irq(q->queue_lock);
|
||||||
return;
|
return;
|
||||||
|
|
||||||
plug_device:
|
plug_device:
|
||||||
if (!elv_queue_empty(orig_drive->queue))
|
spin_unlock_irq(&hwgroup->lock);
|
||||||
blk_plug_device(orig_drive->queue);
|
spin_lock_irq(q->queue_lock);
|
||||||
|
|
||||||
|
if (!elv_queue_empty(q))
|
||||||
|
blk_plug_device(q);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -957,6 +866,17 @@ out:
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void ide_plug_device(ide_drive_t *drive)
|
||||||
|
{
|
||||||
|
struct request_queue *q = drive->queue;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
spin_lock_irqsave(q->queue_lock, flags);
|
||||||
|
if (!elv_queue_empty(q))
|
||||||
|
blk_plug_device(q);
|
||||||
|
spin_unlock_irqrestore(q->queue_lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* ide_timer_expiry - handle lack of an IDE interrupt
|
* ide_timer_expiry - handle lack of an IDE interrupt
|
||||||
* @data: timer callback magic (hwgroup)
|
* @data: timer callback magic (hwgroup)
|
||||||
@ -974,10 +894,12 @@ out:
|
|||||||
void ide_timer_expiry (unsigned long data)
|
void ide_timer_expiry (unsigned long data)
|
||||||
{
|
{
|
||||||
ide_hwgroup_t *hwgroup = (ide_hwgroup_t *) data;
|
ide_hwgroup_t *hwgroup = (ide_hwgroup_t *) data;
|
||||||
|
ide_drive_t *uninitialized_var(drive);
|
||||||
ide_handler_t *handler;
|
ide_handler_t *handler;
|
||||||
ide_expiry_t *expiry;
|
ide_expiry_t *expiry;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
unsigned long wait = -1;
|
unsigned long wait = -1;
|
||||||
|
int plug_device = 0;
|
||||||
|
|
||||||
spin_lock_irqsave(&hwgroup->lock, flags);
|
spin_lock_irqsave(&hwgroup->lock, flags);
|
||||||
|
|
||||||
@ -989,12 +911,8 @@ void ide_timer_expiry (unsigned long data)
|
|||||||
* or we were "sleeping" to give other devices a chance.
|
* or we were "sleeping" to give other devices a chance.
|
||||||
* Either way, we don't really want to complain about anything.
|
* Either way, we don't really want to complain about anything.
|
||||||
*/
|
*/
|
||||||
if (hwgroup->sleeping) {
|
|
||||||
hwgroup->sleeping = 0;
|
|
||||||
ide_unlock_hwgroup(hwgroup);
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
ide_drive_t *drive = hwgroup->drive;
|
drive = hwgroup->drive;
|
||||||
if (!drive) {
|
if (!drive) {
|
||||||
printk(KERN_ERR "ide_timer_expiry: hwgroup->drive was NULL\n");
|
printk(KERN_ERR "ide_timer_expiry: hwgroup->drive was NULL\n");
|
||||||
hwgroup->handler = NULL;
|
hwgroup->handler = NULL;
|
||||||
@ -1042,17 +960,18 @@ void ide_timer_expiry (unsigned long data)
|
|||||||
ide_error(drive, "irq timeout",
|
ide_error(drive, "irq timeout",
|
||||||
hwif->tp_ops->read_status(hwif));
|
hwif->tp_ops->read_status(hwif));
|
||||||
}
|
}
|
||||||
drive->service_time = jiffies - drive->service_start;
|
|
||||||
spin_lock_irq(&hwgroup->lock);
|
spin_lock_irq(&hwgroup->lock);
|
||||||
enable_irq(hwif->irq);
|
enable_irq(hwif->irq);
|
||||||
if (startstop == ide_stopped) {
|
if (startstop == ide_stopped) {
|
||||||
ide_unlock_hwgroup(hwgroup);
|
ide_unlock_hwgroup(hwgroup);
|
||||||
if (!elv_queue_empty(drive->queue))
|
plug_device = 1;
|
||||||
blk_plug_device(drive->queue);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&hwgroup->lock, flags);
|
spin_unlock_irqrestore(&hwgroup->lock, flags);
|
||||||
|
|
||||||
|
if (plug_device)
|
||||||
|
ide_plug_device(drive);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -1146,10 +1065,11 @@ irqreturn_t ide_intr (int irq, void *dev_id)
|
|||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
ide_hwgroup_t *hwgroup = (ide_hwgroup_t *)dev_id;
|
ide_hwgroup_t *hwgroup = (ide_hwgroup_t *)dev_id;
|
||||||
ide_hwif_t *hwif = hwgroup->hwif;
|
ide_hwif_t *hwif = hwgroup->hwif;
|
||||||
ide_drive_t *drive;
|
ide_drive_t *uninitialized_var(drive);
|
||||||
ide_handler_t *handler;
|
ide_handler_t *handler;
|
||||||
ide_startstop_t startstop;
|
ide_startstop_t startstop;
|
||||||
irqreturn_t irq_ret = IRQ_NONE;
|
irqreturn_t irq_ret = IRQ_NONE;
|
||||||
|
int plug_device = 0;
|
||||||
|
|
||||||
spin_lock_irqsave(&hwgroup->lock, flags);
|
spin_lock_irqsave(&hwgroup->lock, flags);
|
||||||
|
|
||||||
@ -1236,12 +1156,10 @@ irqreturn_t ide_intr (int irq, void *dev_id)
|
|||||||
* same irq as is currently being serviced here, and Linux
|
* same irq as is currently being serviced here, and Linux
|
||||||
* won't allow another of the same (on any CPU) until we return.
|
* won't allow another of the same (on any CPU) until we return.
|
||||||
*/
|
*/
|
||||||
drive->service_time = jiffies - drive->service_start;
|
|
||||||
if (startstop == ide_stopped) {
|
if (startstop == ide_stopped) {
|
||||||
if (hwgroup->handler == NULL) { /* paranoia */
|
if (hwgroup->handler == NULL) { /* paranoia */
|
||||||
ide_unlock_hwgroup(hwgroup);
|
ide_unlock_hwgroup(hwgroup);
|
||||||
if (!elv_queue_empty(drive->queue))
|
plug_device = 1;
|
||||||
blk_plug_device(drive->queue);
|
|
||||||
} else
|
} else
|
||||||
printk(KERN_ERR "%s: %s: huh? expected NULL handler "
|
printk(KERN_ERR "%s: %s: huh? expected NULL handler "
|
||||||
"on exit\n", __func__, drive->name);
|
"on exit\n", __func__, drive->name);
|
||||||
@ -1250,6 +1168,10 @@ out_handled:
|
|||||||
irq_ret = IRQ_HANDLED;
|
irq_ret = IRQ_HANDLED;
|
||||||
out:
|
out:
|
||||||
spin_unlock_irqrestore(&hwgroup->lock, flags);
|
spin_unlock_irqrestore(&hwgroup->lock, flags);
|
||||||
|
|
||||||
|
if (plug_device)
|
||||||
|
ide_plug_device(drive);
|
||||||
|
|
||||||
return irq_ret;
|
return irq_ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -16,16 +16,19 @@ static void issue_park_cmd(ide_drive_t *drive, unsigned long timeout)
|
|||||||
spin_lock_irq(&hwgroup->lock);
|
spin_lock_irq(&hwgroup->lock);
|
||||||
if (drive->dev_flags & IDE_DFLAG_PARKED) {
|
if (drive->dev_flags & IDE_DFLAG_PARKED) {
|
||||||
int reset_timer = time_before(timeout, drive->sleep);
|
int reset_timer = time_before(timeout, drive->sleep);
|
||||||
|
int start_queue = 0;
|
||||||
|
|
||||||
drive->sleep = timeout;
|
drive->sleep = timeout;
|
||||||
wake_up_all(&ide_park_wq);
|
wake_up_all(&ide_park_wq);
|
||||||
if (reset_timer && hwgroup->sleeping &&
|
if (reset_timer && del_timer(&hwgroup->timer))
|
||||||
del_timer(&hwgroup->timer)) {
|
start_queue = 1;
|
||||||
hwgroup->sleeping = 0;
|
|
||||||
ide_unlock_hwgroup(hwgroup);
|
|
||||||
blk_start_queueing(q);
|
|
||||||
}
|
|
||||||
spin_unlock_irq(&hwgroup->lock);
|
spin_unlock_irq(&hwgroup->lock);
|
||||||
|
|
||||||
|
if (start_queue) {
|
||||||
|
spin_lock_irq(q->queue_lock);
|
||||||
|
blk_start_queueing(q);
|
||||||
|
spin_unlock_irq(q->queue_lock);
|
||||||
|
}
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
spin_unlock_irq(&hwgroup->lock);
|
spin_unlock_irq(&hwgroup->lock);
|
||||||
|
@ -881,8 +881,7 @@ static int ide_init_queue(ide_drive_t *drive)
|
|||||||
* do not.
|
* do not.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
q = blk_init_queue_node(do_ide_request, &hwif->hwgroup->lock,
|
q = blk_init_queue_node(do_ide_request, NULL, hwif_to_node(hwif));
|
||||||
hwif_to_node(hwif));
|
|
||||||
if (!q)
|
if (!q)
|
||||||
return 1;
|
return 1;
|
||||||
|
|
||||||
|
@ -603,8 +603,6 @@ struct ide_drive_s {
|
|||||||
unsigned long dev_flags;
|
unsigned long dev_flags;
|
||||||
|
|
||||||
unsigned long sleep; /* sleep until this time */
|
unsigned long sleep; /* sleep until this time */
|
||||||
unsigned long service_start; /* time we started last request */
|
|
||||||
unsigned long service_time; /* service time of last request */
|
|
||||||
unsigned long timeout; /* max time to wait for irq */
|
unsigned long timeout; /* max time to wait for irq */
|
||||||
|
|
||||||
special_t special; /* special action flags */
|
special_t special; /* special action flags */
|
||||||
@ -872,8 +870,6 @@ typedef struct hwgroup_s {
|
|||||||
|
|
||||||
/* BOOL: protects all fields below */
|
/* BOOL: protects all fields below */
|
||||||
volatile int busy;
|
volatile int busy;
|
||||||
/* BOOL: wake us up on timer expiry */
|
|
||||||
unsigned int sleeping : 1;
|
|
||||||
/* BOOL: polling active & poll_timeout field valid */
|
/* BOOL: polling active & poll_timeout field valid */
|
||||||
unsigned int polling : 1;
|
unsigned int polling : 1;
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user