Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next
[linux-2.6-block.git] / drivers / hv / channel_mgmt.c
index 90d6545564378c35d55fcb7f71a635ca5cfc32d5..ed9350d42764e06bb38eacfdf8756ae8ac43b2ba 100644 (file)
@@ -149,6 +149,7 @@ static struct vmbus_channel *alloc_channel(void)
        spin_lock_init(&channel->sc_lock);
 
        INIT_LIST_HEAD(&channel->sc_list);
+       INIT_LIST_HEAD(&channel->percpu_list);
 
        channel->controlwq = create_workqueue("hv_vmbus_ctl");
        if (!channel->controlwq) {
@@ -188,7 +189,20 @@ static void free_channel(struct vmbus_channel *channel)
        queue_work(vmbus_connection.work_queue, &channel->work);
 }
 
+static void percpu_channel_enq(void *arg)
+{
+       struct vmbus_channel *channel = arg;
+       int cpu = smp_processor_id();
+
+       list_add_tail(&channel->percpu_list, &hv_context.percpu_list[cpu]);
+}
 
+static void percpu_channel_deq(void *arg)
+{
+       struct vmbus_channel *channel = arg;
+
+       list_del(&channel->percpu_list);
+}
 
 /*
  * vmbus_process_rescind_offer -
@@ -210,6 +224,12 @@ static void vmbus_process_rescind_offer(struct work_struct *work)
        msg.header.msgtype = CHANNELMSG_RELID_RELEASED;
        vmbus_post_msg(&msg, sizeof(struct vmbus_channel_relid_released));
 
+       if (channel->target_cpu != smp_processor_id())
+               smp_call_function_single(channel->target_cpu,
+                                        percpu_channel_deq, channel, true);
+       else
+               percpu_channel_deq(channel);
+
        if (channel->primary_channel == NULL) {
                spin_lock_irqsave(&vmbus_connection.channel_lock, flags);
                list_del(&channel->listentry);
@@ -245,6 +265,7 @@ static void vmbus_process_offer(struct work_struct *work)
                                                        work);
        struct vmbus_channel *channel;
        bool fnew = true;
+       bool enq = false;
        int ret;
        unsigned long flags;
 
@@ -264,12 +285,22 @@ static void vmbus_process_offer(struct work_struct *work)
                }
        }
 
-       if (fnew)
+       if (fnew) {
                list_add_tail(&newchannel->listentry,
                              &vmbus_connection.chn_list);
+               enq = true;
+       }
 
        spin_unlock_irqrestore(&vmbus_connection.channel_lock, flags);
 
+       if (enq) {
+               if (newchannel->target_cpu != smp_processor_id())
+                       smp_call_function_single(newchannel->target_cpu,
+                                                percpu_channel_enq,
+                                                newchannel, true);
+               else
+                       percpu_channel_enq(newchannel);
+       }
        if (!fnew) {
                /*
                 * Check to see if this is a sub-channel.
@@ -282,6 +313,14 @@ static void vmbus_process_offer(struct work_struct *work)
                        spin_lock_irqsave(&channel->sc_lock, flags);
                        list_add_tail(&newchannel->sc_list, &channel->sc_list);
                        spin_unlock_irqrestore(&channel->sc_lock, flags);
+
+                       if (newchannel->target_cpu != smp_processor_id())
+                               smp_call_function_single(newchannel->target_cpu,
+                                                        percpu_channel_enq,
+                                                        newchannel, true);
+                       else
+                               percpu_channel_enq(newchannel);
+
                        newchannel->state = CHANNEL_OPEN_STATE;
                        if (channel->sc_creation_callback != NULL)
                                channel->sc_creation_callback(newchannel);
@@ -365,7 +404,7 @@ static u32  next_vp;
  * performance critical channels (IDE, SCSI and Network) will be uniformly
  * distributed across all available CPUs.
  */
-static u32 get_vp_index(const uuid_le *type_guid)
+static void init_vp_index(struct vmbus_channel *channel, const uuid_le *type_guid)
 {
        u32 cur_cpu;
        int i;
@@ -387,10 +426,13 @@ static u32 get_vp_index(const uuid_le *type_guid)
                 * Also if the channel is not a performance critical
                 * channel, bind it to cpu 0.
                 */
-               return 0;
+               channel->target_cpu = 0;
+               channel->target_vp = 0;
+               return;
        }
        cur_cpu = (++next_vp % max_cpus);
-       return hv_context.vp_index[cur_cpu];
+       channel->target_cpu = cur_cpu;
+       channel->target_vp = hv_context.vp_index[cur_cpu];
 }
 
 /*
@@ -438,7 +480,7 @@ static void vmbus_onoffer(struct vmbus_channel_message_header *hdr)
                                offer->connection_id;
        }
 
-       newchannel->target_vp = get_vp_index(&offer->offer.if_type);
+       init_vp_index(newchannel, &offer->offer.if_type);
 
        memcpy(&newchannel->offermsg, offer,
               sizeof(struct vmbus_channel_offer_channel));