Drivers: hv: vmbus: Replace the per-CPU channel lists with a global array of channels
authorAndrea Parri (Microsoft) <parri.andrea@gmail.com>
Mon, 6 Apr 2020 00:15:06 +0000 (02:15 +0200)
committerWei Liu <wei.liu@kernel.org>
Thu, 23 Apr 2020 13:17:11 +0000 (13:17 +0000)
When Hyper-V sends an interrupt to the guest, the guest has to figure
out which channel the interrupt is associated with.  Hyper-V sets a bit
in a memory page that is shared with the guest, indicating a particular
"relid" that the interrupt is associated with.  The current Linux code
then uses a set of per-CPU linked lists to map a given "relid" to a
pointer to a channel structure.

This design introduces a synchronization problem if the CPU that Hyper-V
will interrupt for a certain channel is changed.  If the interrupt comes
on the "old CPU" and the channel was already moved to the per-CPU list
of the "new CPU", then the relid -> channel mapping will fail and the
interrupt is dropped.  Similarly, if the interrupt comes on the new CPU
but the channel was not moved to the per-CPU list of the new CPU, then
the mapping will fail and the interrupt is dropped.

Relids are integers ranging from 0 to 2047.  The mapping from relids to
channel structures can be done by setting up an array with 2048 entries,
each entry being a pointer to a channel structure (hence total size ~16K
bytes, which is not a problem).  The array is global, so there are no
per-CPU linked lists to update.  The array can be searched and updated
by loading from/storing to the array at the specified index.  With no
per-CPU data structures, the above mentioned synchronization problem is
avoided and the relid2channel() function gets simpler.

Suggested-by: Michael Kelley <mikelley@microsoft.com>
Signed-off-by: Andrea Parri (Microsoft) <parri.andrea@gmail.com>
Link: https://lore.kernel.org/r/20200406001514.19876-4-parri.andrea@gmail.com
Reviewed-by: Michael Kelley <mikelley@microsoft.com>
Signed-off-by: Wei Liu <wei.liu@kernel.org>
drivers/hv/channel_mgmt.c
drivers/hv/connection.c
drivers/hv/hv.c
drivers/hv/hyperv_vmbus.h
drivers/hv/vmbus_drv.c
include/linux/hyperv.h

index 0ced32ef2715a54a6bcabc91b61037fdc71fa99c..bbd0eee4362f67ba5898bc29648277a2863834ce 100644 (file)
@@ -319,7 +319,6 @@ static struct vmbus_channel *alloc_channel(void)
        init_completion(&channel->rescind_event);
 
        INIT_LIST_HEAD(&channel->sc_list);
-       INIT_LIST_HEAD(&channel->percpu_list);
 
        tasklet_init(&channel->callback_event,
                     vmbus_on_event, (unsigned long)channel);
@@ -340,23 +339,49 @@ static void free_channel(struct vmbus_channel *channel)
        kobject_put(&channel->kobj);
 }
 
-static void percpu_channel_enq(void *arg)
+void vmbus_channel_map_relid(struct vmbus_channel *channel)
 {
-       struct vmbus_channel *channel = arg;
-       struct hv_per_cpu_context *hv_cpu
-               = this_cpu_ptr(hv_context.cpu_context);
-
-       list_add_tail_rcu(&channel->percpu_list, &hv_cpu->chan_list);
+       if (WARN_ON(channel->offermsg.child_relid >= MAX_CHANNEL_RELIDS))
+               return;
+       /*
+        * The mapping of the channel's relid is visible from the CPUs that
+        * execute vmbus_chan_sched() by the time that vmbus_chan_sched() will
+        * execute:
+        *
+        *  (a) In the "normal (i.e., not resuming from hibernation)" path,
+        *      the full barrier in smp_store_mb() guarantees that the store
+        *      is propagated to all CPUs before the add_channel_work work
+        *      is queued.  In turn, add_channel_work is queued before the
+        *      channel's ring buffer is allocated/initialized and the
+        *      OPENCHANNEL message for the channel is sent in vmbus_open().
+        *      Hyper-V won't start sending the interrupts for the channel
+        *      before the OPENCHANNEL message is acked.  The memory barrier
+        *      in vmbus_chan_sched() -> sync_test_and_clear_bit() ensures
+        *      that vmbus_chan_sched() must find the channel's relid in
+        *      recv_int_page before retrieving the channel pointer from the
+        *      array of channels.
+        *
+        *  (b) In the "resuming from hibernation" path, the smp_store_mb()
+        *      guarantees that the store is propagated to all CPUs before
+        *      the VMBus connection is marked as ready for the resume event
+        *      (cf. check_ready_for_resume_event()).  The interrupt handler
+        *      of the VMBus driver and vmbus_chan_sched() can not run before
+        *      vmbus_bus_resume() has completed execution (cf. resume_noirq).
+        */
+       smp_store_mb(
+               vmbus_connection.channels[channel->offermsg.child_relid],
+               channel);
 }
 
-static void percpu_channel_deq(void *arg)
+void vmbus_channel_unmap_relid(struct vmbus_channel *channel)
 {
-       struct vmbus_channel *channel = arg;
-
-       list_del_rcu(&channel->percpu_list);
+       if (WARN_ON(channel->offermsg.child_relid >= MAX_CHANNEL_RELIDS))
+               return;
+       WRITE_ONCE(
+               vmbus_connection.channels[channel->offermsg.child_relid],
+               NULL);
 }
 
-
 static void vmbus_release_relid(u32 relid)
 {
        struct vmbus_channel_relid_released msg;
@@ -376,17 +401,25 @@ void hv_process_channel_removal(struct vmbus_channel *channel)
        struct vmbus_channel *primary_channel;
        unsigned long flags;
 
-       BUG_ON(!mutex_is_locked(&vmbus_connection.channel_mutex));
+       lockdep_assert_held(&vmbus_connection.channel_mutex);
        BUG_ON(!channel->rescind);
 
-       if (channel->target_cpu != get_cpu()) {
-               put_cpu();
-               smp_call_function_single(channel->target_cpu,
-                                        percpu_channel_deq, channel, true);
-       } else {
-               percpu_channel_deq(channel);
-               put_cpu();
-       }
+       /*
+        * hv_process_channel_removal() could find INVALID_RELID only for
+        * hv_sock channels.  See the inline comments in vmbus_onoffer().
+        */
+       WARN_ON(channel->offermsg.child_relid == INVALID_RELID &&
+               !is_hvsock_channel(channel));
+
+       /*
+        * Upon suspend, an in-use hv_sock channel is removed from the array of
+        * channels and the relid is invalidated.  After hibernation, when the
+        * user-space appplication destroys the channel, it's unnecessary and
+        * unsafe to remove the channel from the array of channels.  See also
+        * the inline comments before the call of vmbus_release_relid() below.
+        */
+       if (channel->offermsg.child_relid != INVALID_RELID)
+               vmbus_channel_unmap_relid(channel);
 
        if (channel->primary_channel == NULL) {
                list_del(&channel->listentry);
@@ -447,16 +480,6 @@ static void vmbus_add_channel_work(struct work_struct *work)
 
        init_vp_index(newchannel, dev_type);
 
-       if (newchannel->target_cpu != get_cpu()) {
-               put_cpu();
-               smp_call_function_single(newchannel->target_cpu,
-                                        percpu_channel_enq,
-                                        newchannel, true);
-       } else {
-               percpu_channel_enq(newchannel);
-               put_cpu();
-       }
-
        /*
         * This state is used to indicate a successful open
         * so that when we do close the channel normally, we
@@ -523,17 +546,10 @@ err_deq_chan:
                spin_unlock_irqrestore(&primary_channel->lock, flags);
        }
 
-       mutex_unlock(&vmbus_connection.channel_mutex);
+       /* vmbus_process_offer() has mapped the channel. */
+       vmbus_channel_unmap_relid(newchannel);
 
-       if (newchannel->target_cpu != get_cpu()) {
-               put_cpu();
-               smp_call_function_single(newchannel->target_cpu,
-                                        percpu_channel_deq,
-                                        newchannel, true);
-       } else {
-               percpu_channel_deq(newchannel);
-               put_cpu();
-       }
+       mutex_unlock(&vmbus_connection.channel_mutex);
 
        vmbus_release_relid(newchannel->offermsg.child_relid);
 
@@ -599,6 +615,8 @@ static void vmbus_process_offer(struct vmbus_channel *newchannel)
                spin_unlock_irqrestore(&channel->lock, flags);
        }
 
+       vmbus_channel_map_relid(newchannel);
+
        mutex_unlock(&vmbus_connection.channel_mutex);
 
        /*
@@ -940,8 +958,6 @@ static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
        oldchannel = find_primary_channel_by_offer(offer);
 
        if (oldchannel != NULL) {
-               atomic_dec(&vmbus_connection.offer_in_progress);
-
                /*
                 * We're resuming from hibernation: all the sub-channel and
                 * hv_sock channels we had before the hibernation should have
@@ -949,36 +965,65 @@ static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
                 * primary channel that we had before the hibernation.
                 */
 
+               /*
+                * { Initially: channel relid = INVALID_RELID,
+                *              channels[valid_relid] = NULL }
+                *
+                * CPU1                                 CPU2
+                *
+                * [vmbus_onoffer()]                    [vmbus_device_release()]
+                *
+                * LOCK channel_mutex                   LOCK channel_mutex
+                * STORE channel relid = valid_relid    LOAD r1 = channel relid
+                * MAP_RELID channel                    if (r1 != INVALID_RELID)
+                * UNLOCK channel_mutex                   UNMAP_RELID channel
+                *                                      UNLOCK channel_mutex
+                *
+                * Forbids: r1 == valid_relid &&
+                *              channels[valid_relid] == channel
+                *
+                * Note.  r1 can be INVALID_RELID only for an hv_sock channel.
+                * None of the hv_sock channels which were present before the
+                * suspend are re-offered upon the resume.  See the WARN_ON()
+                * in hv_process_channel_removal().
+                */
+               mutex_lock(&vmbus_connection.channel_mutex);
+
+               atomic_dec(&vmbus_connection.offer_in_progress);
+
                WARN_ON(oldchannel->offermsg.child_relid != INVALID_RELID);
                /* Fix up the relid. */
                oldchannel->offermsg.child_relid = offer->child_relid;
 
                offer_sz = sizeof(*offer);
-               if (memcmp(offer, &oldchannel->offermsg, offer_sz) == 0) {
-                       check_ready_for_resume_event();
-                       return;
+               if (memcmp(offer, &oldchannel->offermsg, offer_sz) != 0) {
+                       /*
+                        * This is not an error, since the host can also change
+                        * the other field(s) of the offer, e.g. on WS RS5
+                        * (Build 17763), the offer->connection_id of the
+                        * Mellanox VF vmbus device can change when the host
+                        * reoffers the device upon resume.
+                        */
+                       pr_debug("vmbus offer changed: relid=%d\n",
+                                offer->child_relid);
+
+                       print_hex_dump_debug("Old vmbus offer: ",
+                                            DUMP_PREFIX_OFFSET, 16, 4,
+                                            &oldchannel->offermsg, offer_sz,
+                                            false);
+                       print_hex_dump_debug("New vmbus offer: ",
+                                            DUMP_PREFIX_OFFSET, 16, 4,
+                                            offer, offer_sz, false);
+
+                       /* Fix up the old channel. */
+                       vmbus_setup_channel_state(oldchannel, offer);
                }
 
-               /*
-                * This is not an error, since the host can also change the
-                * other field(s) of the offer, e.g. on WS RS5 (Build 17763),
-                * the offer->connection_id of the Mellanox VF vmbus device
-                * can change when the host reoffers the device upon resume.
-                */
-               pr_debug("vmbus offer changed: relid=%d\n",
-                        offer->child_relid);
-
-               print_hex_dump_debug("Old vmbus offer: ", DUMP_PREFIX_OFFSET,
-                                    16, 4, &oldchannel->offermsg, offer_sz,
-                                    false);
-               print_hex_dump_debug("New vmbus offer: ", DUMP_PREFIX_OFFSET,
-                                    16, 4, offer, offer_sz, false);
-
-               /* Fix up the old channel. */
-               vmbus_setup_channel_state(oldchannel, offer);
-
+               /* Add the channel back to the array of channels. */
+               vmbus_channel_map_relid(oldchannel);
                check_ready_for_resume_event();
 
+               mutex_unlock(&vmbus_connection.channel_mutex);
                return;
        }
 
@@ -1036,14 +1081,14 @@ static void vmbus_onoffer_rescind(struct vmbus_channel_message_header *hdr)
         *
         * CPU1                         CPU2
         *
-        * [vmbus_process_offer()]      [vmbus_onoffer_rescind()]
+        * [vmbus_onoffer()]            [vmbus_onoffer_rescind()]
         *
         * LOCK channel_mutex           WAIT_ON offer_in_progress == 0
         * DECREMENT offer_in_progress  LOCK channel_mutex
-        * INSERT chn_list              SEARCH chn_list
+        * STORE channels[]             LOAD channels[]
         * UNLOCK channel_mutex         UNLOCK channel_mutex
         *
-        * Forbids: CPU2's SEARCH from *not* seeing CPU1's INSERT
+        * Forbids: CPU2's LOAD from *not* seeing CPU1's STORE
         */
 
        while (atomic_read(&vmbus_connection.offer_in_progress) != 0) {
index f4bd306d2cef94bb4a15a1dd813bb132b58ad86e..11170d9a2e1a533b6c3a9c1bf93dd2662c92981a 100644 (file)
@@ -248,6 +248,14 @@ int vmbus_connect(void)
        pr_info("Vmbus version:%d.%d\n",
                version >> 16, version & 0xFFFF);
 
+       vmbus_connection.channels = kcalloc(MAX_CHANNEL_RELIDS,
+                                           sizeof(struct vmbus_channel *),
+                                           GFP_KERNEL);
+       if (vmbus_connection.channels == NULL) {
+               ret = -ENOMEM;
+               goto cleanup;
+       }
+
        kfree(msginfo);
        return 0;
 
@@ -295,33 +303,9 @@ void vmbus_disconnect(void)
  */
 struct vmbus_channel *relid2channel(u32 relid)
 {
-       struct vmbus_channel *channel;
-       struct vmbus_channel *found_channel  = NULL;
-       struct list_head *cur, *tmp;
-       struct vmbus_channel *cur_sc;
-
-       BUG_ON(!mutex_is_locked(&vmbus_connection.channel_mutex));
-
-       list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
-               if (channel->offermsg.child_relid == relid) {
-                       found_channel = channel;
-                       break;
-               } else if (!list_empty(&channel->sc_list)) {
-                       /*
-                        * Deal with sub-channels.
-                        */
-                       list_for_each_safe(cur, tmp, &channel->sc_list) {
-                               cur_sc = list_entry(cur, struct vmbus_channel,
-                                                       sc_list);
-                               if (cur_sc->offermsg.child_relid == relid) {
-                                       found_channel = cur_sc;
-                                       break;
-                               }
-                       }
-               }
-       }
-
-       return found_channel;
+       if (WARN_ON(relid >= MAX_CHANNEL_RELIDS))
+               return NULL;
+       return READ_ONCE(vmbus_connection.channels[relid]);
 }
 
 /*
index e2b331045464082bad2abb4b5e14411c38f33b92..17bf1f229152b75898b2169c7f4439fd182093c9 100644 (file)
@@ -117,8 +117,6 @@ int hv_synic_alloc(void)
                        pr_err("Unable to allocate post msg page\n");
                        goto err;
                }
-
-               INIT_LIST_HEAD(&hv_cpu->chan_list);
        }
 
        return 0;
index 000740d053a2e1df6ed12defaf1bf0af333c6a8b..3edf993b0fd9b0240ee10dc5f1c173d9e65fe154 100644 (file)
@@ -132,12 +132,6 @@ struct hv_per_cpu_context {
         * basis.
         */
        struct tasklet_struct msg_dpc;
-
-       /*
-        * To optimize the mapping of relid to channel, maintain
-        * per-cpu list of the channels based on their CPU affinity.
-        */
-       struct list_head chan_list;
 };
 
 struct hv_context {
@@ -202,6 +196,8 @@ int hv_ringbuffer_read(struct vmbus_channel *channel,
 /* TODO: Need to make this configurable */
 #define MAX_NUM_CHANNELS_SUPPORTED     256
 
+#define MAX_CHANNEL_RELIDS                                     \
+       max(MAX_NUM_CHANNELS_SUPPORTED, HV_EVENT_FLAGS_COUNT)
 
 enum vmbus_connect_state {
        DISCONNECTED,
@@ -251,6 +247,9 @@ struct vmbus_connection {
        struct list_head chn_list;
        struct mutex channel_mutex;
 
+       /* Array of channels */
+       struct vmbus_channel **channels;
+
        /*
         * An offer message is handled first on the work_queue, and then
         * is further handled on handle_primary_chan_wq or
@@ -338,6 +337,9 @@ int vmbus_add_channel_kobj(struct hv_device *device_obj,
 
 void vmbus_remove_channel_attr_group(struct vmbus_channel *channel);
 
+void vmbus_channel_map_relid(struct vmbus_channel *channel);
+void vmbus_channel_unmap_relid(struct vmbus_channel *channel);
+
 struct vmbus_channel *relid2channel(u32 relid);
 
 void vmbus_free_channels(void);
index 1f3e69d18d3563e91262aad9ec1ee39b9e783fb8..21a01bca7fcdb19a056e6e5437a2510e5e2b9e2a 100644 (file)
@@ -1252,33 +1252,39 @@ static void vmbus_chan_sched(struct hv_per_cpu_context *hv_cpu)
                if (relid == 0)
                        continue;
 
+               /*
+                * Pairs with the kfree_rcu() in vmbus_chan_release().
+                * Guarantees that the channel data structure doesn't
+                * get freed while the channel pointer below is being
+                * dereferenced.
+                */
                rcu_read_lock();
 
                /* Find channel based on relid */
-               list_for_each_entry_rcu(channel, &hv_cpu->chan_list, percpu_list) {
-                       if (channel->offermsg.child_relid != relid)
-                               continue;
+               channel = relid2channel(relid);
+               if (channel == NULL)
+                       goto sched_unlock_rcu;
 
-                       if (channel->rescind)
-                               continue;
+               if (channel->rescind)
+                       goto sched_unlock_rcu;
 
-                       trace_vmbus_chan_sched(channel);
+               trace_vmbus_chan_sched(channel);
 
-                       ++channel->interrupts;
+               ++channel->interrupts;
 
-                       switch (channel->callback_mode) {
-                       case HV_CALL_ISR:
-                               vmbus_channel_isr(channel);
-                               break;
+               switch (channel->callback_mode) {
+               case HV_CALL_ISR:
+                       vmbus_channel_isr(channel);
+                       break;
 
-                       case HV_CALL_BATCHED:
-                               hv_begin_read(&channel->inbound);
-                               /* fallthrough */
-                       case HV_CALL_DIRECT:
-                               tasklet_schedule(&channel->callback_event);
-                       }
+               case HV_CALL_BATCHED:
+                       hv_begin_read(&channel->inbound);
+                       fallthrough;
+               case HV_CALL_DIRECT:
+                       tasklet_schedule(&channel->callback_event);
                }
 
+sched_unlock_rcu:
                rcu_read_unlock();
        }
 }
@@ -2264,9 +2270,12 @@ static int vmbus_bus_suspend(struct device *dev)
 
        list_for_each_entry(channel, &vmbus_connection.chn_list, listentry) {
                /*
-                * Invalidate the field. Upon resume, vmbus_onoffer() will fix
-                * up the field, and the other fields (if necessary).
+                * Remove the channel from the array of channels and invalidate
+                * the channel's relid.  Upon resume, vmbus_onoffer() will fix
+                * up the relid (and other fields, if necessary) and add the
+                * channel back to the array.
                 */
+               vmbus_channel_unmap_relid(channel);
                channel->offermsg.child_relid = INVALID_RELID;
 
                if (is_hvsock_channel(channel)) {
@@ -2502,6 +2511,7 @@ static void __exit vmbus_exit(void)
        hv_debug_rm_all_dir();
 
        vmbus_free_channels();
+       kfree(vmbus_connection.channels);
 
        if (ms_hyperv.misc_features & HV_FEATURE_GUEST_CRASH_MSR_AVAILABLE) {
                kmsg_dump_unregister(&hv_kmsg_dumper);
index cbd24f4e68d15a8c8585c9346857196f65f07858..f5b3f008c55afd987eec555e3494a2ab9bc2db98 100644 (file)
@@ -854,11 +854,6 @@ struct vmbus_channel {
         * Support per-channel state for use by vmbus drivers.
         */
        void *per_channel_state;
-       /*
-        * To support per-cpu lookup mapping of relid to channel,
-        * link up channels based on their CPU affinity.
-        */
-       struct list_head percpu_list;
 
        /*
         * Defer freeing channel until after all cpu's have