Drivers: hv: vmbus: Replace the per-CPU channel lists with a global array of channels
When Hyper-V sends an interrupt to the guest, the guest has to figure out which channel the interrupt is associated with. Hyper-V sets a bit in a memory page that is shared with the guest, indicating a particular "relid" that the interrupt is associated with. The current Linux code then uses a set of per-CPU linked lists to map a given "relid" to a pointer to a channel structure. This design introduces a synchronization problem if the CPU that Hyper-V will interrupt for a certain channel is changed. If the interrupt comes on the "old CPU" and the channel was already moved to the per-CPU list of the "new CPU", then the relid -> channel mapping will fail and the interrupt is dropped. Similarly, if the interrupt comes on the new CPU but the channel was not moved to the per-CPU list of the new CPU, then the mapping will fail and the interrupt is dropped. Relids are integers ranging from 0 to 2047. The mapping from relids to channel structures can be done by setting up an array with 2048 entries, each entry being a pointer to a channel structure (hence total size ~16K bytes, which is not a problem). The array is global, so there are no per-CPU linked lists to update. The array can be searched and updated by loading from/storing to the array at the specified index. With no per-CPU data structures, the above mentioned synchronization problem is avoided and the relid2channel() function gets simpler. Suggested-by: Michael Kelley <mikelley@microsoft.com> Signed-off-by: Andrea Parri (Microsoft) <parri.andrea@gmail.com> Link: https://lore.kernel.org/r/20200406001514.19876-4-parri.andrea@gmail.com Reviewed-by: Michael Kelley <mikelley@microsoft.com> Signed-off-by: Wei Liu <wei.liu@kernel.org>
This commit is contained in:
parent
b9fa1b8797
commit
8b6a877c06
@ -319,7 +319,6 @@ static struct vmbus_channel *alloc_channel(void)
|
|||||||
init_completion(&channel->rescind_event);
|
init_completion(&channel->rescind_event);
|
||||||
|
|
||||||
INIT_LIST_HEAD(&channel->sc_list);
|
INIT_LIST_HEAD(&channel->sc_list);
|
||||||
INIT_LIST_HEAD(&channel->percpu_list);
|
|
||||||
|
|
||||||
tasklet_init(&channel->callback_event,
|
tasklet_init(&channel->callback_event,
|
||||||
vmbus_on_event, (unsigned long)channel);
|
vmbus_on_event, (unsigned long)channel);
|
||||||
@ -340,23 +339,49 @@ static void free_channel(struct vmbus_channel *channel)
|
|||||||
kobject_put(&channel->kobj);
|
kobject_put(&channel->kobj);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void percpu_channel_enq(void *arg)
|
void vmbus_channel_map_relid(struct vmbus_channel *channel)
|
||||||
{
|
{
|
||||||
struct vmbus_channel *channel = arg;
|
if (WARN_ON(channel->offermsg.child_relid >= MAX_CHANNEL_RELIDS))
|
||||||
struct hv_per_cpu_context *hv_cpu
|
return;
|
||||||
= this_cpu_ptr(hv_context.cpu_context);
|
/*
|
||||||
|
* The mapping of the channel's relid is visible from the CPUs that
|
||||||
list_add_tail_rcu(&channel->percpu_list, &hv_cpu->chan_list);
|
* execute vmbus_chan_sched() by the time that vmbus_chan_sched() will
|
||||||
|
* execute:
|
||||||
|
*
|
||||||
|
* (a) In the "normal (i.e., not resuming from hibernation)" path,
|
||||||
|
* the full barrier in smp_store_mb() guarantees that the store
|
||||||
|
* is propagated to all CPUs before the add_channel_work work
|
||||||
|
* is queued. In turn, add_channel_work is queued before the
|
||||||
|
* channel's ring buffer is allocated/initialized and the
|
||||||
|
* OPENCHANNEL message for the channel is sent in vmbus_open().
|
||||||
|
* Hyper-V won't start sending the interrupts for the channel
|
||||||
|
* before the OPENCHANNEL message is acked. The memory barrier
|
||||||
|
* in vmbus_chan_sched() -> sync_test_and_clear_bit() ensures
|
||||||
|
* that vmbus_chan_sched() must find the channel's relid in
|
||||||
|
* recv_int_page before retrieving the channel pointer from the
|
||||||
|
* array of channels.
|
||||||
|
*
|
||||||
|
* (b) In the "resuming from hibernation" path, the smp_store_mb()
|
||||||
|
* guarantees that the store is propagated to all CPUs before
|
||||||
|
* the VMBus connection is marked as ready for the resume event
|
||||||
|
* (cf. check_ready_for_resume_event()). The interrupt handler
|
||||||
|
* of the VMBus driver and vmbus_chan_sched() can not run before
|
||||||
|
* vmbus_bus_resume() has completed execution (cf. resume_noirq).
|
||||||
|
*/
|
||||||
|
smp_store_mb(
|
||||||
|
vmbus_connection.channels[channel->offermsg.child_relid],
|
||||||
|
channel);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void percpu_channel_deq(void *arg)
|
void vmbus_channel_unmap_relid(struct vmbus_channel *channel)
|
||||||
{
|
{
|
||||||
struct vmbus_channel *channel = arg;
|
if (WARN_ON(channel->offermsg.child_relid >= MAX_CHANNEL_RELIDS))
|
||||||
|
return;
|
||||||
list_del_rcu(&channel->percpu_list);
|
WRITE_ONCE(
|
||||||
|
vmbus_connection.channels[channel->offermsg.child_relid],
|
||||||
|
NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static void vmbus_release_relid(u32 relid)
|
static void vmbus_release_relid(u32 relid)
|
||||||
{
|
{
|
||||||
struct vmbus_channel_relid_released msg;
|
struct vmbus_channel_relid_released msg;
|
||||||
@ -376,17 +401,25 @@ void hv_process_channel_removal(struct vmbus_channel *channel)
|
|||||||
struct vmbus_channel *primary_channel;
|
struct vmbus_channel *primary_channel;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
BUG_ON(!mutex_is_locked(&vmbus_connection.channel_mutex));
|
lockdep_assert_held(&vmbus_connection.channel_mutex);
|
||||||
BUG_ON(!channel->rescind);
|
BUG_ON(!channel->rescind);
|
||||||
|
|
||||||
if (channel->target_cpu != get_cpu()) {
|
/*
|
||||||
put_cpu();
|
* hv_process_channel_removal() could find INVALID_RELID only for
|
||||||
smp_call_function_single(channel->target_cpu,
|
* hv_sock channels. See the inline comments in vmbus_onoffer().
|
||||||
percpu_channel_deq, channel, true);
|
*/
|
||||||
} else {
|
WARN_ON(channel->offermsg.child_relid == INVALID_RELID &&
|
||||||
percpu_channel_deq(channel);
|
!is_hvsock_channel(channel));
|
||||||
put_cpu();
|
|
||||||
}
|
/*
|
||||||
|
* Upon suspend, an in-use hv_sock channel is removed from the array of
|
||||||
|
* channels and the relid is invalidated. After hibernation, when the
|
||||||
|
* user-space appplication destroys the channel, it's unnecessary and
|
||||||
|
* unsafe to remove the channel from the array of channels. See also
|
||||||
|
* the inline comments before the call of vmbus_release_relid() below.
|
||||||
|
*/
|
||||||
|
if (channel->offermsg.child_relid != INVALID_RELID)
|
||||||
|
vmbus_channel_unmap_relid(channel);
|
||||||
|
|
||||||
if (channel->primary_channel == NULL) {
|
if (channel->primary_channel == NULL) {
|
||||||
list_del(&channel->listentry);
|
list_del(&channel->listentry);
|
||||||
@ -447,16 +480,6 @@ static void vmbus_add_channel_work(struct work_struct *work)
|
|||||||
|
|
||||||
init_vp_index(newchannel, dev_type);
|
init_vp_index(newchannel, dev_type);
|
||||||
|
|
||||||
if (newchannel->target_cpu != get_cpu()) {
|
|
||||||
put_cpu();
|
|
||||||
smp_call_function_single(newchannel->target_cpu,
|
|
||||||
percpu_channel_enq,
|
|
||||||
newchannel, true);
|
|
||||||
} else {
|
|
||||||
percpu_channel_enq(newchannel);
|
|
||||||
put_cpu();
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This state is used to indicate a successful open
|
* This state is used to indicate a successful open
|
||||||
* so that when we do close the channel normally, we
|
* so that when we do close the channel normally, we
|
||||||
@ -523,17 +546,10 @@ err_deq_chan:
|
|||||||
spin_unlock_irqrestore(&primary_channel->lock, flags);
|
spin_unlock_irqrestore(&primary_channel->lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
mutex_unlock(&vmbus_connection.channel_mutex);
|
/* vmbus_process_offer() has mapped the channel. */
|
||||||
|
vmbus_channel_unmap_relid(newchannel);
|
||||||
|
|
||||||
if (newchannel->target_cpu != get_cpu()) {
|
mutex_unlock(&vmbus_connection.channel_mutex);
|
||||||
put_cpu();
|
|
||||||
smp_call_function_single(newchannel->target_cpu,
|
|
||||||
percpu_channel_deq,
|
|
||||||
newchannel, true);
|
|
||||||
} else {
|
|
||||||
percpu_channel_deq(newchannel);
|
|
||||||
put_cpu();
|
|
||||||
}
|
|
||||||
|
|
||||||
vmbus_release_relid(newchannel->offermsg.child_relid);
|
vmbus_release_relid(newchannel->offermsg.child_relid);
|
||||||
|
|
||||||
@ -599,6 +615,8 @@ static void vmbus_process_offer(struct vmbus_channel *newchannel)
|
|||||||
spin_unlock_irqrestore(&channel->lock, flags);
|
spin_unlock_irqrestore(&channel->lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
vmbus_channel_map_relid(newchannel);
|
||||||
|
|
||||||
mutex_unlock(&vmbus_connection.channel_mutex);
|
mutex_unlock(&vmbus_connection.channel_mutex);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -940,8 +958,6 @@ static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
|
|||||||
oldchannel = find_primary_channel_by_offer(offer);
|
oldchannel = find_primary_channel_by_offer(offer);
|
||||||
|
|
||||||
if (oldchannel != NULL) {
|
if (oldchannel != NULL) {
|
||||||
atomic_dec(&vmbus_connection.offer_in_progress);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We're resuming from hibernation: all the sub-channel and
|
* We're resuming from hibernation: all the sub-channel and
|
||||||
* hv_sock channels we had before the hibernation should have
|
* hv_sock channels we had before the hibernation should have
|
||||||
@ -949,36 +965,65 @@ static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
|
|||||||
* primary channel that we had before the hibernation.
|
* primary channel that we had before the hibernation.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
/*
|
||||||
|
* { Initially: channel relid = INVALID_RELID,
|
||||||
|
* channels[valid_relid] = NULL }
|
||||||
|
*
|
||||||
|
* CPU1 CPU2
|
||||||
|
*
|
||||||
|
* [vmbus_onoffer()] [vmbus_device_release()]
|
||||||
|
*
|
||||||
|
* LOCK channel_mutex LOCK channel_mutex
|
||||||
|
* STORE channel relid = valid_relid LOAD r1 = channel relid
|
||||||
|
* MAP_RELID channel if (r1 != INVALID_RELID)
|
||||||
|
* UNLOCK channel_mutex UNMAP_RELID channel
|
||||||
|
* UNLOCK channel_mutex
|
||||||
|
*
|
||||||
|
* Forbids: r1 == valid_relid &&
|
||||||
|
* channels[valid_relid] == channel
|
||||||
|
*
|
||||||
|
* Note. r1 can be INVALID_RELID only for an hv_sock channel.
|
||||||
|
* None of the hv_sock channels which were present before the
|
||||||
|
* suspend are re-offered upon the resume. See the WARN_ON()
|
||||||
|
* in hv_process_channel_removal().
|
||||||
|
*/
|
||||||
|
mutex_lock(&vmbus_connection.channel_mutex);
|
||||||
|
|
||||||
|
atomic_dec(&vmbus_connection.offer_in_progress);
|
||||||
|
|
||||||
WARN_ON(oldchannel->offermsg.child_relid != INVALID_RELID);
|
WARN_ON(oldchannel->offermsg.child_relid != INVALID_RELID);
|
||||||
/* Fix up the relid. */
|
/* Fix up the relid. */
|
||||||
oldchannel->offermsg.child_relid = offer->child_relid;
|
oldchannel->offermsg.child_relid = offer->child_relid;
|
||||||
|
|
||||||
offer_sz = sizeof(*offer);
|
offer_sz = sizeof(*offer);
|
||||||
if (memcmp(offer, &oldchannel->offermsg, offer_sz) == 0) {
|
if (memcmp(offer, &oldchannel->offermsg, offer_sz) != 0) {
|
||||||
check_ready_for_resume_event();
|
/*
|
||||||
return;
|
* This is not an error, since the host can also change
|
||||||
|
* the other field(s) of the offer, e.g. on WS RS5
|
||||||
|
* (Build 17763), the offer->connection_id of the
|
||||||
|
* Mellanox VF vmbus device can change when the host
|
||||||
|
* reoffers the device upon resume.
|
||||||
|
*/
|
||||||
|
pr_debug("vmbus offer changed: relid=%d\n",
|
||||||
|
offer->child_relid);
|
||||||
|
|
||||||
|
print_hex_dump_debug("Old vmbus offer: ",
|
||||||
|
DUMP_PREFIX_OFFSET, 16, 4,
|
||||||
|
&oldchannel->offermsg, offer_sz,
|
||||||
|
false);
|
||||||
|
print_hex_dump_debug("New vmbus offer: ",
|
||||||
|
DUMP_PREFIX_OFFSET, 16, 4,
|
||||||
|
offer, offer_sz, false);
|
||||||
|
|
||||||
|
/* Fix up the old channel. */
|
||||||
|
vmbus_setup_channel_state(oldchannel, offer);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Add the channel back to the array of channels. */
|
||||||
* This is not an error, since the host can also change the
|
vmbus_channel_map_relid(oldchannel);
|
||||||
* other field(s) of the offer, e.g. on WS RS5 (Build 17763),
|
|
||||||
* the offer->connection_id of the Mellanox VF vmbus device
|
|
||||||
* can change when the host reoffers the device upon resume.
|
|
||||||
*/
|
|
||||||
pr_debug("vmbus offer changed: relid=%d\n",
|
|
||||||
offer->child_relid);
|
|
||||||
|
|
||||||
print_hex_dump_debug("Old vmbus offer: ", DUMP_PREFIX_OFFSET,
|
|
||||||
16, 4, &oldchannel->offermsg, offer_sz,
|
|
||||||
false);
|
|
||||||
print_hex_dump_debug("New vmbus offer: ", DUMP_PREFIX_OFFSET,
|
|
||||||
16, 4, offer, offer_sz, false);
|
|
||||||
|
|
||||||
/* Fix up the old channel. */
|
|
||||||
vmbus_setup_channel_state(oldchannel, offer);
|
|
||||||
|
|
||||||
check_ready_for_resume_event();
|
check_ready_for_resume_event();
|
||||||
|
|
||||||
|
mutex_unlock(&vmbus_connection.channel_mutex);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1036,14 +1081,14 @@ static void vmbus_onoffer_rescind(struct vmbus_channel_message_header *hdr)
|
|||||||
*
|
*
|
||||||
* CPU1 CPU2
|
* CPU1 CPU2
|
||||||
*
|
*
|
||||||
* [vmbus_process_offer()] [vmbus_onoffer_rescind()]
|
* [vmbus_onoffer()] [vmbus_onoffer_rescind()]
|
||||||
*
|
*
|
||||||
* LOCK channel_mutex WAIT_ON offer_in_progress == 0
|
* LOCK channel_mutex WAIT_ON offer_in_progress == 0
|
||||||
* DECREMENT offer_in_progress LOCK channel_mutex
|
* DECREMENT offer_in_progress LOCK channel_mutex
|
||||||
* INSERT chn_list SEARCH chn_list
|
* STORE channels[] LOAD channels[]
|
||||||
* UNLOCK channel_mutex UNLOCK channel_mutex
|
* UNLOCK channel_mutex UNLOCK channel_mutex
|
||||||
*
|
*
|
||||||
* Forbids: CPU2's SEARCH from *not* seeing CPU1's INSERT
|
* Forbids: CPU2's LOAD from *not* seeing CPU1's STORE
|
||||||
*/
|
*/
|
||||||
|
|
||||||
while (atomic_read(&vmbus_connection.offer_in_progress) != 0) {
|
while (atomic_read(&vmbus_connection.offer_in_progress) != 0) {
|
||||||
|
@ -248,6 +248,14 @@ int vmbus_connect(void)
|
|||||||
pr_info("Vmbus version:%d.%d\n",
|
pr_info("Vmbus version:%d.%d\n",
|
||||||
version >> 16, version & 0xFFFF);
|
version >> 16, version & 0xFFFF);
|
||||||
|
|
||||||
|
vmbus_connection.channels = kcalloc(MAX_CHANNEL_RELIDS,
|
||||||
|
sizeof(struct vmbus_channel *),
|
||||||
|
GFP_KERNEL);
|
||||||
|
if (vmbus_connection.channels == NULL) {
|
||||||
|
ret = -ENOMEM;
|
||||||
|
goto cleanup;
|
||||||
|
}
|
||||||
|
|
||||||
kfree(msginfo);
|
kfree(msginfo);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
@ -295,33 +303,9 @@ void vmbus_disconnect(void)
|
|||||||
*/
|
*/
|
||||||
struct vmbus_channel *relid2channel(u32 relid)
|
struct vmbus_channel *relid2channel(u32 relid)
|
||||||
{
|
{
|
||||||
struct vmbus_channel *channel;
|
if (WARN_ON(relid >= MAX_CHANNEL_RELIDS))
|
||||||
struct vmbus_channel *found_channel = NULL;
|
return NULL;
|
||||||
struct list_head *cur, *tmp;
|
return READ_ONCE(vmbus_connection.channels[relid]);
|
||||||
struct vmbus_channel *cur_sc;
|
|
||||||
|
|
||||||
BUG_ON(!mutex_is_locked(&vmbus_connection.channel_mutex));
|
|
||||||
|
|
||||||
list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
|
|
||||||
if (channel->offermsg.child_relid == relid) {
|
|
||||||
found_channel = channel;
|
|
||||||
break;
|
|
||||||
} else if (!list_empty(&channel->sc_list)) {
|
|
||||||
/*
|
|
||||||
* Deal with sub-channels.
|
|
||||||
*/
|
|
||||||
list_for_each_safe(cur, tmp, &channel->sc_list) {
|
|
||||||
cur_sc = list_entry(cur, struct vmbus_channel,
|
|
||||||
sc_list);
|
|
||||||
if (cur_sc->offermsg.child_relid == relid) {
|
|
||||||
found_channel = cur_sc;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return found_channel;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -117,8 +117,6 @@ int hv_synic_alloc(void)
|
|||||||
pr_err("Unable to allocate post msg page\n");
|
pr_err("Unable to allocate post msg page\n");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
INIT_LIST_HEAD(&hv_cpu->chan_list);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -132,12 +132,6 @@ struct hv_per_cpu_context {
|
|||||||
* basis.
|
* basis.
|
||||||
*/
|
*/
|
||||||
struct tasklet_struct msg_dpc;
|
struct tasklet_struct msg_dpc;
|
||||||
|
|
||||||
/*
|
|
||||||
* To optimize the mapping of relid to channel, maintain
|
|
||||||
* per-cpu list of the channels based on their CPU affinity.
|
|
||||||
*/
|
|
||||||
struct list_head chan_list;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
struct hv_context {
|
struct hv_context {
|
||||||
@ -202,6 +196,8 @@ int hv_ringbuffer_read(struct vmbus_channel *channel,
|
|||||||
/* TODO: Need to make this configurable */
|
/* TODO: Need to make this configurable */
|
||||||
#define MAX_NUM_CHANNELS_SUPPORTED 256
|
#define MAX_NUM_CHANNELS_SUPPORTED 256
|
||||||
|
|
||||||
|
#define MAX_CHANNEL_RELIDS \
|
||||||
|
max(MAX_NUM_CHANNELS_SUPPORTED, HV_EVENT_FLAGS_COUNT)
|
||||||
|
|
||||||
enum vmbus_connect_state {
|
enum vmbus_connect_state {
|
||||||
DISCONNECTED,
|
DISCONNECTED,
|
||||||
@ -251,6 +247,9 @@ struct vmbus_connection {
|
|||||||
struct list_head chn_list;
|
struct list_head chn_list;
|
||||||
struct mutex channel_mutex;
|
struct mutex channel_mutex;
|
||||||
|
|
||||||
|
/* Array of channels */
|
||||||
|
struct vmbus_channel **channels;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* An offer message is handled first on the work_queue, and then
|
* An offer message is handled first on the work_queue, and then
|
||||||
* is further handled on handle_primary_chan_wq or
|
* is further handled on handle_primary_chan_wq or
|
||||||
@ -338,6 +337,9 @@ int vmbus_add_channel_kobj(struct hv_device *device_obj,
|
|||||||
|
|
||||||
void vmbus_remove_channel_attr_group(struct vmbus_channel *channel);
|
void vmbus_remove_channel_attr_group(struct vmbus_channel *channel);
|
||||||
|
|
||||||
|
void vmbus_channel_map_relid(struct vmbus_channel *channel);
|
||||||
|
void vmbus_channel_unmap_relid(struct vmbus_channel *channel);
|
||||||
|
|
||||||
struct vmbus_channel *relid2channel(u32 relid);
|
struct vmbus_channel *relid2channel(u32 relid);
|
||||||
|
|
||||||
void vmbus_free_channels(void);
|
void vmbus_free_channels(void);
|
||||||
|
@ -1252,33 +1252,39 @@ static void vmbus_chan_sched(struct hv_per_cpu_context *hv_cpu)
|
|||||||
if (relid == 0)
|
if (relid == 0)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Pairs with the kfree_rcu() in vmbus_chan_release().
|
||||||
|
* Guarantees that the channel data structure doesn't
|
||||||
|
* get freed while the channel pointer below is being
|
||||||
|
* dereferenced.
|
||||||
|
*/
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
|
|
||||||
/* Find channel based on relid */
|
/* Find channel based on relid */
|
||||||
list_for_each_entry_rcu(channel, &hv_cpu->chan_list, percpu_list) {
|
channel = relid2channel(relid);
|
||||||
if (channel->offermsg.child_relid != relid)
|
if (channel == NULL)
|
||||||
continue;
|
goto sched_unlock_rcu;
|
||||||
|
|
||||||
if (channel->rescind)
|
if (channel->rescind)
|
||||||
continue;
|
goto sched_unlock_rcu;
|
||||||
|
|
||||||
trace_vmbus_chan_sched(channel);
|
trace_vmbus_chan_sched(channel);
|
||||||
|
|
||||||
++channel->interrupts;
|
++channel->interrupts;
|
||||||
|
|
||||||
switch (channel->callback_mode) {
|
switch (channel->callback_mode) {
|
||||||
case HV_CALL_ISR:
|
case HV_CALL_ISR:
|
||||||
vmbus_channel_isr(channel);
|
vmbus_channel_isr(channel);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case HV_CALL_BATCHED:
|
case HV_CALL_BATCHED:
|
||||||
hv_begin_read(&channel->inbound);
|
hv_begin_read(&channel->inbound);
|
||||||
/* fallthrough */
|
fallthrough;
|
||||||
case HV_CALL_DIRECT:
|
case HV_CALL_DIRECT:
|
||||||
tasklet_schedule(&channel->callback_event);
|
tasklet_schedule(&channel->callback_event);
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
sched_unlock_rcu:
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -2264,9 +2270,12 @@ static int vmbus_bus_suspend(struct device *dev)
|
|||||||
|
|
||||||
list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
|
list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
|
||||||
/*
|
/*
|
||||||
* Invalidate the field. Upon resume, vmbus_onoffer() will fix
|
* Remove the channel from the array of channels and invalidate
|
||||||
* up the field, and the other fields (if necessary).
|
* the channel's relid. Upon resume, vmbus_onoffer() will fix
|
||||||
|
* up the relid (and other fields, if necessary) and add the
|
||||||
|
* channel back to the array.
|
||||||
*/
|
*/
|
||||||
|
vmbus_channel_unmap_relid(channel);
|
||||||
channel->offermsg.child_relid = INVALID_RELID;
|
channel->offermsg.child_relid = INVALID_RELID;
|
||||||
|
|
||||||
if (is_hvsock_channel(channel)) {
|
if (is_hvsock_channel(channel)) {
|
||||||
@ -2502,6 +2511,7 @@ static void __exit vmbus_exit(void)
|
|||||||
hv_debug_rm_all_dir();
|
hv_debug_rm_all_dir();
|
||||||
|
|
||||||
vmbus_free_channels();
|
vmbus_free_channels();
|
||||||
|
kfree(vmbus_connection.channels);
|
||||||
|
|
||||||
if (ms_hyperv.misc_features & HV_FEATURE_GUEST_CRASH_MSR_AVAILABLE) {
|
if (ms_hyperv.misc_features & HV_FEATURE_GUEST_CRASH_MSR_AVAILABLE) {
|
||||||
kmsg_dump_unregister(&hv_kmsg_dumper);
|
kmsg_dump_unregister(&hv_kmsg_dumper);
|
||||||
|
@ -854,11 +854,6 @@ struct vmbus_channel {
|
|||||||
* Support per-channel state for use by vmbus drivers.
|
* Support per-channel state for use by vmbus drivers.
|
||||||
*/
|
*/
|
||||||
void *per_channel_state;
|
void *per_channel_state;
|
||||||
/*
|
|
||||||
* To support per-cpu lookup mapping of relid to channel,
|
|
||||||
* link up channels based on their CPU affinity.
|
|
||||||
*/
|
|
||||||
struct list_head percpu_list;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Defer freeing channel until after all cpu's have
|
* Defer freeing channel until after all cpu's have
|
||||||
|
Loading…
Reference in New Issue
Block a user