Drivers: hv: hv_balloon: refuse to balloon below the floor
[linux-block.git] / drivers / hv / vmbus_drv.c
CommitLineData
3e7ee490 1/*
3e7ee490
HJ
2 * Copyright (c) 2009, Microsoft Corporation.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
11 * more details.
12 *
13 * You should have received a copy of the GNU General Public License along with
14 * this program; if not, write to the Free Software Foundation, Inc., 59 Temple
15 * Place - Suite 330, Boston, MA 02111-1307 USA.
16 *
17 * Authors:
18 * Haiyang Zhang <haiyangz@microsoft.com>
19 * Hank Janssen <hjanssen@microsoft.com>
b0069f43 20 * K. Y. Srinivasan <kys@microsoft.com>
52e5c1ce 21 *
3e7ee490 22 */
0a46618d
HJ
23#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
24
3e7ee490
HJ
25#include <linux/init.h>
26#include <linux/module.h>
27#include <linux/device.h>
3e7ee490
HJ
28#include <linux/interrupt.h>
29#include <linux/sysctl.h>
5a0e3ad6 30#include <linux/slab.h>
b0069f43 31#include <linux/acpi.h>
8b5d6d3b 32#include <linux/completion.h>
46a97191 33#include <linux/hyperv.h>
b0209501 34#include <linux/kernel_stat.h>
4061ed9e 35#include <linux/clockchips.h>
e513229b 36#include <linux/cpu.h>
407dd164 37#include <asm/hyperv.h>
1f94ea81 38#include <asm/hypervisor.h>
302a3c0f 39#include <asm/mshyperv.h>
0f2a6619 40#include "hyperv_vmbus.h"
3e7ee490 41
607c1a11 42static struct acpi_device *hv_acpi_dev;
1168ac22 43
59c0e4f0 44static struct tasklet_struct msg_dpc;
71a6655d 45static struct completion probe_event;
b0069f43 46static int irq;
98db4335 47
90eedf0c
GH
48struct resource hyperv_mmio = {
49 .name = "hyperv mmio",
50 .flags = IORESOURCE_MEM,
51};
52EXPORT_SYMBOL_GPL(hyperv_mmio);
98db4335 53
cf6a2eac
S
54static int vmbus_exists(void)
55{
56 if (hv_acpi_dev == NULL)
57 return -ENODEV;
58
59 return 0;
60}
61
fd776ba9
OH
62#define VMBUS_ALIAS_LEN ((sizeof((struct hv_vmbus_device_id *)0)->guid) * 2)
63static void print_alias_name(struct hv_device *hv_dev, char *alias_name)
64{
65 int i;
66 for (i = 0; i < VMBUS_ALIAS_LEN; i += 2)
67 sprintf(&alias_name[i], "%02x", hv_dev->dev_type.b[i/2]);
68}
69
76c52bbe
GKH
70static u8 channel_monitor_group(struct vmbus_channel *channel)
71{
72 return (u8)channel->offermsg.monitorid / 32;
73}
74
75static u8 channel_monitor_offset(struct vmbus_channel *channel)
76{
77 return (u8)channel->offermsg.monitorid % 32;
78}
79
80static u32 channel_pending(struct vmbus_channel *channel,
81 struct hv_monitor_page *monitor_page)
82{
83 u8 monitor_group = channel_monitor_group(channel);
84 return monitor_page->trigger_group[monitor_group].pending;
85}
86
1cee272b
GKH
87static u32 channel_latency(struct vmbus_channel *channel,
88 struct hv_monitor_page *monitor_page)
89{
90 u8 monitor_group = channel_monitor_group(channel);
91 u8 monitor_offset = channel_monitor_offset(channel);
92 return monitor_page->latency[monitor_group][monitor_offset];
93}
94
4947c745
GKH
95static u32 channel_conn_id(struct vmbus_channel *channel,
96 struct hv_monitor_page *monitor_page)
97{
98 u8 monitor_group = channel_monitor_group(channel);
99 u8 monitor_offset = channel_monitor_offset(channel);
100 return monitor_page->parameter[monitor_group][monitor_offset].connectionid.u.id;
101}
102
03f3a910
GKH
103static ssize_t id_show(struct device *dev, struct device_attribute *dev_attr,
104 char *buf)
105{
106 struct hv_device *hv_dev = device_to_hv_device(dev);
107
108 if (!hv_dev->channel)
109 return -ENODEV;
110 return sprintf(buf, "%d\n", hv_dev->channel->offermsg.child_relid);
111}
112static DEVICE_ATTR_RO(id);
113
a8fb5f3d
GKH
114static ssize_t state_show(struct device *dev, struct device_attribute *dev_attr,
115 char *buf)
116{
117 struct hv_device *hv_dev = device_to_hv_device(dev);
118
119 if (!hv_dev->channel)
120 return -ENODEV;
121 return sprintf(buf, "%d\n", hv_dev->channel->state);
122}
123static DEVICE_ATTR_RO(state);
124
5ffd00e2
GKH
125static ssize_t monitor_id_show(struct device *dev,
126 struct device_attribute *dev_attr, char *buf)
127{
128 struct hv_device *hv_dev = device_to_hv_device(dev);
129
130 if (!hv_dev->channel)
131 return -ENODEV;
132 return sprintf(buf, "%d\n", hv_dev->channel->offermsg.monitorid);
133}
134static DEVICE_ATTR_RO(monitor_id);
135
68234c04
GKH
136static ssize_t class_id_show(struct device *dev,
137 struct device_attribute *dev_attr, char *buf)
138{
139 struct hv_device *hv_dev = device_to_hv_device(dev);
140
141 if (!hv_dev->channel)
142 return -ENODEV;
143 return sprintf(buf, "{%pUl}\n",
144 hv_dev->channel->offermsg.offer.if_type.b);
145}
146static DEVICE_ATTR_RO(class_id);
147
7c55e1d0
GKH
148static ssize_t device_id_show(struct device *dev,
149 struct device_attribute *dev_attr, char *buf)
150{
151 struct hv_device *hv_dev = device_to_hv_device(dev);
152
153 if (!hv_dev->channel)
154 return -ENODEV;
155 return sprintf(buf, "{%pUl}\n",
156 hv_dev->channel->offermsg.offer.if_instance.b);
157}
158static DEVICE_ATTR_RO(device_id);
159
647fa371
GKH
160static ssize_t modalias_show(struct device *dev,
161 struct device_attribute *dev_attr, char *buf)
162{
163 struct hv_device *hv_dev = device_to_hv_device(dev);
164 char alias_name[VMBUS_ALIAS_LEN + 1];
165
166 print_alias_name(hv_dev, alias_name);
167 return sprintf(buf, "vmbus:%s\n", alias_name);
168}
169static DEVICE_ATTR_RO(modalias);
170
76c52bbe
GKH
171static ssize_t server_monitor_pending_show(struct device *dev,
172 struct device_attribute *dev_attr,
173 char *buf)
174{
175 struct hv_device *hv_dev = device_to_hv_device(dev);
176
177 if (!hv_dev->channel)
178 return -ENODEV;
179 return sprintf(buf, "%d\n",
180 channel_pending(hv_dev->channel,
181 vmbus_connection.monitor_pages[1]));
182}
183static DEVICE_ATTR_RO(server_monitor_pending);
184
185static ssize_t client_monitor_pending_show(struct device *dev,
186 struct device_attribute *dev_attr,
187 char *buf)
188{
189 struct hv_device *hv_dev = device_to_hv_device(dev);
190
191 if (!hv_dev->channel)
192 return -ENODEV;
193 return sprintf(buf, "%d\n",
194 channel_pending(hv_dev->channel,
195 vmbus_connection.monitor_pages[1]));
196}
197static DEVICE_ATTR_RO(client_monitor_pending);
68234c04 198
1cee272b
GKH
199static ssize_t server_monitor_latency_show(struct device *dev,
200 struct device_attribute *dev_attr,
201 char *buf)
202{
203 struct hv_device *hv_dev = device_to_hv_device(dev);
204
205 if (!hv_dev->channel)
206 return -ENODEV;
207 return sprintf(buf, "%d\n",
208 channel_latency(hv_dev->channel,
209 vmbus_connection.monitor_pages[0]));
210}
211static DEVICE_ATTR_RO(server_monitor_latency);
212
213static ssize_t client_monitor_latency_show(struct device *dev,
214 struct device_attribute *dev_attr,
215 char *buf)
216{
217 struct hv_device *hv_dev = device_to_hv_device(dev);
218
219 if (!hv_dev->channel)
220 return -ENODEV;
221 return sprintf(buf, "%d\n",
222 channel_latency(hv_dev->channel,
223 vmbus_connection.monitor_pages[1]));
224}
225static DEVICE_ATTR_RO(client_monitor_latency);
226
4947c745
GKH
227static ssize_t server_monitor_conn_id_show(struct device *dev,
228 struct device_attribute *dev_attr,
229 char *buf)
230{
231 struct hv_device *hv_dev = device_to_hv_device(dev);
232
233 if (!hv_dev->channel)
234 return -ENODEV;
235 return sprintf(buf, "%d\n",
236 channel_conn_id(hv_dev->channel,
237 vmbus_connection.monitor_pages[0]));
238}
239static DEVICE_ATTR_RO(server_monitor_conn_id);
240
241static ssize_t client_monitor_conn_id_show(struct device *dev,
242 struct device_attribute *dev_attr,
243 char *buf)
244{
245 struct hv_device *hv_dev = device_to_hv_device(dev);
246
247 if (!hv_dev->channel)
248 return -ENODEV;
249 return sprintf(buf, "%d\n",
250 channel_conn_id(hv_dev->channel,
251 vmbus_connection.monitor_pages[1]));
252}
253static DEVICE_ATTR_RO(client_monitor_conn_id);
254
98f4c651
GKH
255static ssize_t out_intr_mask_show(struct device *dev,
256 struct device_attribute *dev_attr, char *buf)
257{
258 struct hv_device *hv_dev = device_to_hv_device(dev);
259 struct hv_ring_buffer_debug_info outbound;
260
261 if (!hv_dev->channel)
262 return -ENODEV;
263 hv_ringbuffer_get_debuginfo(&hv_dev->channel->outbound, &outbound);
264 return sprintf(buf, "%d\n", outbound.current_interrupt_mask);
265}
266static DEVICE_ATTR_RO(out_intr_mask);
267
268static ssize_t out_read_index_show(struct device *dev,
269 struct device_attribute *dev_attr, char *buf)
270{
271 struct hv_device *hv_dev = device_to_hv_device(dev);
272 struct hv_ring_buffer_debug_info outbound;
273
274 if (!hv_dev->channel)
275 return -ENODEV;
276 hv_ringbuffer_get_debuginfo(&hv_dev->channel->outbound, &outbound);
277 return sprintf(buf, "%d\n", outbound.current_read_index);
278}
279static DEVICE_ATTR_RO(out_read_index);
280
281static ssize_t out_write_index_show(struct device *dev,
282 struct device_attribute *dev_attr,
283 char *buf)
284{
285 struct hv_device *hv_dev = device_to_hv_device(dev);
286 struct hv_ring_buffer_debug_info outbound;
287
288 if (!hv_dev->channel)
289 return -ENODEV;
290 hv_ringbuffer_get_debuginfo(&hv_dev->channel->outbound, &outbound);
291 return sprintf(buf, "%d\n", outbound.current_write_index);
292}
293static DEVICE_ATTR_RO(out_write_index);
294
295static ssize_t out_read_bytes_avail_show(struct device *dev,
296 struct device_attribute *dev_attr,
297 char *buf)
298{
299 struct hv_device *hv_dev = device_to_hv_device(dev);
300 struct hv_ring_buffer_debug_info outbound;
301
302 if (!hv_dev->channel)
303 return -ENODEV;
304 hv_ringbuffer_get_debuginfo(&hv_dev->channel->outbound, &outbound);
305 return sprintf(buf, "%d\n", outbound.bytes_avail_toread);
306}
307static DEVICE_ATTR_RO(out_read_bytes_avail);
308
309static ssize_t out_write_bytes_avail_show(struct device *dev,
310 struct device_attribute *dev_attr,
311 char *buf)
312{
313 struct hv_device *hv_dev = device_to_hv_device(dev);
314 struct hv_ring_buffer_debug_info outbound;
315
316 if (!hv_dev->channel)
317 return -ENODEV;
318 hv_ringbuffer_get_debuginfo(&hv_dev->channel->outbound, &outbound);
319 return sprintf(buf, "%d\n", outbound.bytes_avail_towrite);
320}
321static DEVICE_ATTR_RO(out_write_bytes_avail);
322
323static ssize_t in_intr_mask_show(struct device *dev,
324 struct device_attribute *dev_attr, char *buf)
325{
326 struct hv_device *hv_dev = device_to_hv_device(dev);
327 struct hv_ring_buffer_debug_info inbound;
328
329 if (!hv_dev->channel)
330 return -ENODEV;
331 hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound);
332 return sprintf(buf, "%d\n", inbound.current_interrupt_mask);
333}
334static DEVICE_ATTR_RO(in_intr_mask);
335
336static ssize_t in_read_index_show(struct device *dev,
337 struct device_attribute *dev_attr, char *buf)
338{
339 struct hv_device *hv_dev = device_to_hv_device(dev);
340 struct hv_ring_buffer_debug_info inbound;
341
342 if (!hv_dev->channel)
343 return -ENODEV;
344 hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound);
345 return sprintf(buf, "%d\n", inbound.current_read_index);
346}
347static DEVICE_ATTR_RO(in_read_index);
348
349static ssize_t in_write_index_show(struct device *dev,
350 struct device_attribute *dev_attr, char *buf)
351{
352 struct hv_device *hv_dev = device_to_hv_device(dev);
353 struct hv_ring_buffer_debug_info inbound;
354
355 if (!hv_dev->channel)
356 return -ENODEV;
357 hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound);
358 return sprintf(buf, "%d\n", inbound.current_write_index);
359}
360static DEVICE_ATTR_RO(in_write_index);
361
362static ssize_t in_read_bytes_avail_show(struct device *dev,
363 struct device_attribute *dev_attr,
364 char *buf)
365{
366 struct hv_device *hv_dev = device_to_hv_device(dev);
367 struct hv_ring_buffer_debug_info inbound;
368
369 if (!hv_dev->channel)
370 return -ENODEV;
371 hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound);
372 return sprintf(buf, "%d\n", inbound.bytes_avail_toread);
373}
374static DEVICE_ATTR_RO(in_read_bytes_avail);
375
376static ssize_t in_write_bytes_avail_show(struct device *dev,
377 struct device_attribute *dev_attr,
378 char *buf)
379{
380 struct hv_device *hv_dev = device_to_hv_device(dev);
381 struct hv_ring_buffer_debug_info inbound;
382
383 if (!hv_dev->channel)
384 return -ENODEV;
385 hv_ringbuffer_get_debuginfo(&hv_dev->channel->inbound, &inbound);
386 return sprintf(buf, "%d\n", inbound.bytes_avail_towrite);
387}
388static DEVICE_ATTR_RO(in_write_bytes_avail);
389
390/* Set up per device attributes in /sys/bus/vmbus/devices/<bus device> */
03f3a910
GKH
391static struct attribute *vmbus_attrs[] = {
392 &dev_attr_id.attr,
a8fb5f3d 393 &dev_attr_state.attr,
5ffd00e2 394 &dev_attr_monitor_id.attr,
68234c04 395 &dev_attr_class_id.attr,
7c55e1d0 396 &dev_attr_device_id.attr,
647fa371 397 &dev_attr_modalias.attr,
76c52bbe
GKH
398 &dev_attr_server_monitor_pending.attr,
399 &dev_attr_client_monitor_pending.attr,
1cee272b
GKH
400 &dev_attr_server_monitor_latency.attr,
401 &dev_attr_client_monitor_latency.attr,
4947c745
GKH
402 &dev_attr_server_monitor_conn_id.attr,
403 &dev_attr_client_monitor_conn_id.attr,
98f4c651
GKH
404 &dev_attr_out_intr_mask.attr,
405 &dev_attr_out_read_index.attr,
406 &dev_attr_out_write_index.attr,
407 &dev_attr_out_read_bytes_avail.attr,
408 &dev_attr_out_write_bytes_avail.attr,
409 &dev_attr_in_intr_mask.attr,
410 &dev_attr_in_read_index.attr,
411 &dev_attr_in_write_index.attr,
412 &dev_attr_in_read_bytes_avail.attr,
413 &dev_attr_in_write_bytes_avail.attr,
03f3a910
GKH
414 NULL,
415};
416ATTRIBUTE_GROUPS(vmbus);
417
adde2487
S
418/*
419 * vmbus_uevent - add uevent for our device
420 *
421 * This routine is invoked when a device is added or removed on the vmbus to
422 * generate a uevent to udev in the userspace. The udev will then look at its
423 * rule and the uevent generated here to load the appropriate driver
0ddda660
S
424 *
425 * The alias string will be of the form vmbus:guid where guid is the string
426 * representation of the device guid (each byte of the guid will be
427 * represented with two hex characters.
adde2487
S
428 */
429static int vmbus_uevent(struct device *device, struct kobj_uevent_env *env)
430{
431 struct hv_device *dev = device_to_hv_device(device);
fd776ba9
OH
432 int ret;
433 char alias_name[VMBUS_ALIAS_LEN + 1];
0ddda660 434
fd776ba9 435 print_alias_name(dev, alias_name);
0ddda660
S
436 ret = add_uevent_var(env, "MODALIAS=vmbus:%s", alias_name);
437 return ret;
adde2487
S
438}
439
1b9d48f2 440static const uuid_le null_guid;
5841a829
S
441
442static inline bool is_null_guid(const __u8 *guid)
443{
444 if (memcmp(guid, &null_guid, sizeof(uuid_le)))
445 return false;
446 return true;
447}
448
3037a7b6
S
449/*
450 * Return a matching hv_vmbus_device_id pointer.
451 * If there is no match, return NULL.
452 */
453static const struct hv_vmbus_device_id *hv_vmbus_get_id(
454 const struct hv_vmbus_device_id *id,
1b9d48f2 455 const __u8 *guid)
3037a7b6
S
456{
457 for (; !is_null_guid(id->guid); id++)
458 if (!memcmp(&id->guid, guid, sizeof(uuid_le)))
459 return id;
460
461 return NULL;
462}
463
464
b7fc147b
S
465
466/*
467 * vmbus_match - Attempt to match the specified device to the specified driver
468 */
469static int vmbus_match(struct device *device, struct device_driver *driver)
470{
b7fc147b 471 struct hv_driver *drv = drv_to_hv_drv(driver);
e8e27047 472 struct hv_device *hv_dev = device_to_hv_device(device);
b7fc147b 473
3037a7b6
S
474 if (hv_vmbus_get_id(drv->id_table, hv_dev->dev_type.b))
475 return 1;
de632a2b 476
5841a829 477 return 0;
b7fc147b
S
478}
479
f1f0d67b
S
480/*
481 * vmbus_probe - Add the new vmbus's child device
482 */
483static int vmbus_probe(struct device *child_device)
484{
485 int ret = 0;
486 struct hv_driver *drv =
487 drv_to_hv_drv(child_device->driver);
9efd21e1 488 struct hv_device *dev = device_to_hv_device(child_device);
84946899 489 const struct hv_vmbus_device_id *dev_id;
f1f0d67b 490
84946899 491 dev_id = hv_vmbus_get_id(drv->id_table, dev->dev_type.b);
9efd21e1 492 if (drv->probe) {
84946899 493 ret = drv->probe(dev, dev_id);
b14a7b30 494 if (ret != 0)
0a46618d
HJ
495 pr_err("probe failed for device %s (%d)\n",
496 dev_name(child_device), ret);
f1f0d67b 497
f1f0d67b 498 } else {
0a46618d
HJ
499 pr_err("probe not set for driver %s\n",
500 dev_name(child_device));
6de925b1 501 ret = -ENODEV;
f1f0d67b
S
502 }
503 return ret;
504}
505
c5dce3db
S
506/*
507 * vmbus_remove - Remove a vmbus device
508 */
509static int vmbus_remove(struct device *child_device)
510{
d15a0301 511 struct hv_driver *drv;
415b023a 512 struct hv_device *dev = device_to_hv_device(child_device);
ed6cfcc5 513 u32 relid = dev->channel->offermsg.child_relid;
c5dce3db 514
d15a0301
S
515 if (child_device->driver) {
516 drv = drv_to_hv_drv(child_device->driver);
517 if (drv->remove)
518 drv->remove(dev);
ed6cfcc5
S
519 else {
520 hv_process_channel_removal(dev->channel, relid);
d15a0301
S
521 pr_err("remove not set for driver %s\n",
522 dev_name(child_device));
ed6cfcc5
S
523 }
524 } else {
525 /*
526 * We don't have a driver for this device; deal with the
527 * rescind message by removing the channel.
528 */
529 hv_process_channel_removal(dev->channel, relid);
d15a0301 530 }
c5dce3db
S
531
532 return 0;
533}
534
eb1bb259
S
535
536/*
537 * vmbus_shutdown - Shutdown a vmbus device
538 */
539static void vmbus_shutdown(struct device *child_device)
540{
541 struct hv_driver *drv;
ca6887fb 542 struct hv_device *dev = device_to_hv_device(child_device);
eb1bb259
S
543
544
545 /* The device may not be attached yet */
546 if (!child_device->driver)
547 return;
548
549 drv = drv_to_hv_drv(child_device->driver);
550
ca6887fb
S
551 if (drv->shutdown)
552 drv->shutdown(dev);
eb1bb259
S
553
554 return;
555}
556
086e7a56
S
557
558/*
559 * vmbus_device_release - Final callback release of the vmbus child device
560 */
561static void vmbus_device_release(struct device *device)
562{
e8e27047 563 struct hv_device *hv_dev = device_to_hv_device(device);
086e7a56 564
e8e27047 565 kfree(hv_dev);
086e7a56
S
566
567}
568
454f18a9 569/* The one and only one */
9adcac5c
S
570static struct bus_type hv_bus = {
571 .name = "vmbus",
572 .match = vmbus_match,
573 .shutdown = vmbus_shutdown,
574 .remove = vmbus_remove,
575 .probe = vmbus_probe,
576 .uevent = vmbus_uevent,
03f3a910 577 .dev_groups = vmbus_groups,
3e7ee490
HJ
578};
579
bf6506f6
TT
580struct onmessage_work_context {
581 struct work_struct work;
582 struct hv_message msg;
583};
584
585static void vmbus_onmessage_work(struct work_struct *work)
586{
587 struct onmessage_work_context *ctx;
588
09a19628
VK
589 /* Do not process messages if we're in DISCONNECTED state */
590 if (vmbus_connection.conn_state == DISCONNECTED)
591 return;
592
bf6506f6
TT
593 ctx = container_of(work, struct onmessage_work_context,
594 work);
595 vmbus_onmessage(&ctx->msg);
596 kfree(ctx);
597}
598
d8a60e00 599static void hv_process_timer_expiration(struct hv_message *msg, int cpu)
4061ed9e
S
600{
601 struct clock_event_device *dev = hv_context.clk_evt[cpu];
602
603 if (dev->event_handler)
604 dev->event_handler(dev);
605
606 msg->header.message_type = HVMSG_NONE;
607
608 /*
609 * Make sure the write to MessageType (ie set to
610 * HVMSG_NONE) happens before we read the
611 * MessagePending and EOMing. Otherwise, the EOMing
612 * will not deliver any more messages since there is
613 * no empty slot
614 */
615 mb();
616
617 if (msg->header.message_flags.msg_pending) {
618 /*
619 * This will cause message queue rescan to
620 * possibly deliver another msg from the
621 * hypervisor
622 */
623 wrmsrl(HV_X64_MSR_EOM, 0);
624 }
625}
626
62c1059d 627static void vmbus_on_msg_dpc(unsigned long data)
36199a99
GKH
628{
629 int cpu = smp_processor_id();
630 void *page_addr = hv_context.synic_message_page[cpu];
631 struct hv_message *msg = (struct hv_message *)page_addr +
632 VMBUS_MESSAGE_SINT;
bf6506f6 633 struct onmessage_work_context *ctx;
36199a99
GKH
634
635 while (1) {
636 if (msg->header.message_type == HVMSG_NONE) {
637 /* no msg */
638 break;
639 } else {
bf6506f6
TT
640 ctx = kmalloc(sizeof(*ctx), GFP_ATOMIC);
641 if (ctx == NULL)
36199a99 642 continue;
bf6506f6
TT
643 INIT_WORK(&ctx->work, vmbus_onmessage_work);
644 memcpy(&ctx->msg, msg, sizeof(*msg));
da9fcb72 645 queue_work(vmbus_connection.work_queue, &ctx->work);
36199a99
GKH
646 }
647
648 msg->header.message_type = HVMSG_NONE;
649
650 /*
651 * Make sure the write to MessageType (ie set to
652 * HVMSG_NONE) happens before we read the
653 * MessagePending and EOMing. Otherwise, the EOMing
654 * will not deliver any more messages since there is
655 * no empty slot
656 */
35848f68 657 mb();
36199a99
GKH
658
659 if (msg->header.message_flags.msg_pending) {
660 /*
661 * This will cause message queue rescan to
662 * possibly deliver another msg from the
663 * hypervisor
664 */
665 wrmsrl(HV_X64_MSR_EOM, 0);
666 }
667 }
668}
669
76d388cd 670static void vmbus_isr(void)
36199a99 671{
36199a99
GKH
672 int cpu = smp_processor_id();
673 void *page_addr;
674 struct hv_message *msg;
675 union hv_synic_event_flags *event;
ae4636e6 676 bool handled = false;
36199a99 677
5ab05951
S
678 page_addr = hv_context.synic_event_page[cpu];
679 if (page_addr == NULL)
76d388cd 680 return;
5ab05951
S
681
682 event = (union hv_synic_event_flags *)page_addr +
683 VMBUS_MESSAGE_SINT;
7341d908
S
684 /*
685 * Check for events before checking for messages. This is the order
686 * in which events and messages are checked in Windows guests on
687 * Hyper-V, and the Windows team suggested we do the same.
688 */
36199a99 689
6552ecd7
S
690 if ((vmbus_proto_version == VERSION_WS2008) ||
691 (vmbus_proto_version == VERSION_WIN7)) {
36199a99 692
6552ecd7
S
693 /* Since we are a child, we only need to check bit 0 */
694 if (sync_test_and_clear_bit(0,
695 (unsigned long *) &event->flags32[0])) {
696 handled = true;
697 }
698 } else {
699 /*
700 * Our host is win8 or above. The signaling mechanism
701 * has changed and we can directly look at the event page.
702 * If bit n is set then we have an interrup on the channel
703 * whose id is n.
704 */
ae4636e6 705 handled = true;
ae4636e6 706 }
793be9c7 707
6552ecd7 708 if (handled)
db11f12a 709 tasklet_schedule(hv_context.event_dpc[cpu]);
6552ecd7
S
710
711
7341d908
S
712 page_addr = hv_context.synic_message_page[cpu];
713 msg = (struct hv_message *)page_addr + VMBUS_MESSAGE_SINT;
714
715 /* Check if there are actual msgs to be processed */
4061ed9e
S
716 if (msg->header.message_type != HVMSG_NONE) {
717 if (msg->header.message_type == HVMSG_TIMER_EXPIRED)
718 hv_process_timer_expiration(msg, cpu);
719 else
720 tasklet_schedule(&msg_dpc);
721 }
793be9c7
S
722}
723
e513229b
VK
724#ifdef CONFIG_HOTPLUG_CPU
725static int hyperv_cpu_disable(void)
726{
727 return -ENOSYS;
728}
729
730static void hv_cpu_hotplug_quirk(bool vmbus_loaded)
731{
732 static void *previous_cpu_disable;
733
734 /*
735 * Offlining a CPU when running on newer hypervisors (WS2012R2, Win8,
736 * ...) is not supported at this moment as channel interrupts are
737 * distributed across all of them.
738 */
739
740 if ((vmbus_proto_version == VERSION_WS2008) ||
741 (vmbus_proto_version == VERSION_WIN7))
742 return;
743
744 if (vmbus_loaded) {
745 previous_cpu_disable = smp_ops.cpu_disable;
746 smp_ops.cpu_disable = hyperv_cpu_disable;
747 pr_notice("CPU offlining is not supported by hypervisor\n");
748 } else if (previous_cpu_disable)
749 smp_ops.cpu_disable = previous_cpu_disable;
750}
751#else
752static void hv_cpu_hotplug_quirk(bool vmbus_loaded)
753{
754}
755#endif
756
3e189519 757/*
90c9960e
GKH
758 * vmbus_bus_init -Main vmbus driver initialization routine.
759 *
760 * Here, we
0686e4f4 761 * - initialize the vmbus driver context
0686e4f4
LL
762 * - invoke the vmbus hv main init routine
763 * - get the irq resource
0686e4f4 764 * - retrieve the channel offers
90c9960e 765 */
9aaa995e 766static int vmbus_bus_init(int irq)
3e7ee490 767{
90c9960e 768 int ret;
3e7ee490 769
6d26e38f
GKH
770 /* Hypervisor initialization...setup hypercall page..etc */
771 ret = hv_init();
90c9960e 772 if (ret != 0) {
0a46618d 773 pr_err("Unable to initialize the hypervisor - 0x%x\n", ret);
d6c1c5de 774 return ret;
3e7ee490
HJ
775 }
776
59c0e4f0 777 tasklet_init(&msg_dpc, vmbus_on_msg_dpc, 0);
3e7ee490 778
9adcac5c 779 ret = bus_register(&hv_bus);
d6c1c5de 780 if (ret)
8b9987e9 781 goto err_cleanup;
3e7ee490 782
76d388cd 783 hv_setup_vmbus_irq(vmbus_isr);
3e7ee490 784
2608fb65
JW
785 ret = hv_synic_alloc();
786 if (ret)
787 goto err_alloc;
800b6902 788 /*
302a3c0f 789 * Initialize the per-cpu interrupt state and
800b6902
S
790 * connect to the host.
791 */
302a3c0f 792 on_each_cpu(hv_synic_init, NULL, 1);
800b6902 793 ret = vmbus_connect();
8b9987e9 794 if (ret)
2608fb65 795 goto err_alloc;
800b6902 796
e513229b 797 hv_cpu_hotplug_quirk(true);
2d6e882b 798 vmbus_request_offers();
8b5d6d3b 799
d6c1c5de 800 return 0;
8b9987e9 801
2608fb65
JW
802err_alloc:
803 hv_synic_free();
76d388cd 804 hv_remove_vmbus_irq();
8b9987e9 805
8b9987e9
S
806 bus_unregister(&hv_bus);
807
808err_cleanup:
809 hv_cleanup();
810
811 return ret;
3e7ee490
HJ
812}
813
90c9960e 814/**
768fa219
GKH
815 * __vmbus_child_driver_register - Register a vmbus's driver
816 * @drv: Pointer to driver structure you want to register
817 * @owner: owner module of the drv
818 * @mod_name: module name string
3e189519
HJ
819 *
820 * Registers the given driver with Linux through the 'driver_register()' call
768fa219 821 * and sets up the hyper-v vmbus handling for this driver.
3e189519
HJ
822 * It will return the state of the 'driver_register()' call.
823 *
90c9960e 824 */
768fa219 825int __vmbus_driver_register(struct hv_driver *hv_driver, struct module *owner, const char *mod_name)
3e7ee490 826{
5d48a1c2 827 int ret;
3e7ee490 828
768fa219 829 pr_info("registering driver %s\n", hv_driver->name);
3e7ee490 830
cf6a2eac
S
831 ret = vmbus_exists();
832 if (ret < 0)
833 return ret;
834
768fa219
GKH
835 hv_driver->driver.name = hv_driver->name;
836 hv_driver->driver.owner = owner;
837 hv_driver->driver.mod_name = mod_name;
838 hv_driver->driver.bus = &hv_bus;
3e7ee490 839
768fa219 840 ret = driver_register(&hv_driver->driver);
3e7ee490 841
5d48a1c2 842 return ret;
3e7ee490 843}
768fa219 844EXPORT_SYMBOL_GPL(__vmbus_driver_register);
3e7ee490 845
90c9960e 846/**
768fa219
GKH
847 * vmbus_driver_unregister() - Unregister a vmbus's driver
848 * @drv: Pointer to driver structure you want to un-register
3e189519 849 *
768fa219
GKH
850 * Un-register the given driver that was previous registered with a call to
851 * vmbus_driver_register()
90c9960e 852 */
768fa219 853void vmbus_driver_unregister(struct hv_driver *hv_driver)
3e7ee490 854{
768fa219 855 pr_info("unregistering driver %s\n", hv_driver->name);
3e7ee490 856
cf6a2eac 857 if (!vmbus_exists())
8f257a14 858 driver_unregister(&hv_driver->driver);
3e7ee490 859}
768fa219 860EXPORT_SYMBOL_GPL(vmbus_driver_unregister);
3e7ee490 861
3e189519 862/*
f2c73011 863 * vmbus_device_create - Creates and registers a new child device
3e189519 864 * on the vmbus.
90c9960e 865 */
1b9d48f2 866struct hv_device *vmbus_device_create(const uuid_le *type,
867 const uuid_le *instance,
868 struct vmbus_channel *channel)
3e7ee490 869{
3d3b5518 870 struct hv_device *child_device_obj;
3e7ee490 871
6bad88da
S
872 child_device_obj = kzalloc(sizeof(struct hv_device), GFP_KERNEL);
873 if (!child_device_obj) {
0a46618d 874 pr_err("Unable to allocate device object for child device\n");
3e7ee490
HJ
875 return NULL;
876 }
877
cae5b843 878 child_device_obj->channel = channel;
358d2ee2 879 memcpy(&child_device_obj->dev_type, type, sizeof(uuid_le));
ca623ad3 880 memcpy(&child_device_obj->dev_instance, instance,
358d2ee2 881 sizeof(uuid_le));
3e7ee490 882
3e7ee490 883
3e7ee490
HJ
884 return child_device_obj;
885}
886
3e189519 887/*
22794281 888 * vmbus_device_register - Register the child device
90c9960e 889 */
22794281 890int vmbus_device_register(struct hv_device *child_device_obj)
3e7ee490 891{
90c9960e 892 int ret = 0;
6bad88da 893
bc63b6f6
VK
894 dev_set_name(&child_device_obj->device, "vmbus_%d",
895 child_device_obj->channel->id);
3e7ee490 896
0bce28b6 897 child_device_obj->device.bus = &hv_bus;
607c1a11 898 child_device_obj->device.parent = &hv_acpi_dev->dev;
6bad88da 899 child_device_obj->device.release = vmbus_device_release;
3e7ee490 900
90c9960e
GKH
901 /*
902 * Register with the LDM. This will kick off the driver/device
903 * binding...which will eventually call vmbus_match() and vmbus_probe()
904 */
6bad88da 905 ret = device_register(&child_device_obj->device);
3e7ee490 906
3e7ee490 907 if (ret)
0a46618d 908 pr_err("Unable to register child device\n");
3e7ee490 909 else
84672369 910 pr_debug("child device %s registered\n",
0a46618d 911 dev_name(&child_device_obj->device));
3e7ee490 912
3e7ee490
HJ
913 return ret;
914}
915
3e189519 916/*
696453ba 917 * vmbus_device_unregister - Remove the specified child device
3e189519 918 * from the vmbus.
90c9960e 919 */
696453ba 920void vmbus_device_unregister(struct hv_device *device_obj)
3e7ee490 921{
84672369
FS
922 pr_debug("child device %s unregistered\n",
923 dev_name(&device_obj->device));
924
90c9960e
GKH
925 /*
926 * Kick off the process of unregistering the device.
927 * This will call vmbus_remove() and eventually vmbus_device_release()
928 */
6bad88da 929 device_unregister(&device_obj->device);
3e7ee490
HJ
930}
931
3e7ee490 932
b0069f43 933/*
90f34535
S
934 * VMBUS is an acpi enumerated device. Get the the information we
935 * need from DSDT.
b0069f43
S
936 */
937
90f34535 938static acpi_status vmbus_walk_resources(struct acpi_resource *res, void *ctx)
b0069f43 939{
90f34535
S
940 switch (res->type) {
941 case ACPI_RESOURCE_TYPE_IRQ:
942 irq = res->data.irq.interrupts[0];
4eb923f8 943 break;
b0069f43 944
90f34535 945 case ACPI_RESOURCE_TYPE_ADDRESS64:
a45de93e
LZ
946 hyperv_mmio.start = res->data.address64.address.minimum;
947 hyperv_mmio.end = res->data.address64.address.maximum;
4eb923f8 948 break;
b0069f43
S
949 }
950
951 return AE_OK;
952}
953
954static int vmbus_acpi_add(struct acpi_device *device)
955{
956 acpi_status result;
90f34535 957 int ret_val = -ENODEV;
b0069f43 958
607c1a11
S
959 hv_acpi_dev = device;
960
0a4425b6 961 result = acpi_walk_resources(device->handle, METHOD_NAME__CRS,
90f34535 962 vmbus_walk_resources, NULL);
b0069f43 963
90f34535
S
964 if (ACPI_FAILURE(result))
965 goto acpi_walk_err;
966 /*
967 * The parent of the vmbus acpi device (Gen2 firmware) is the VMOD that
968 * has the mmio ranges. Get that.
969 */
970 if (device->parent) {
971 result = acpi_walk_resources(device->parent->handle,
972 METHOD_NAME__CRS,
973 vmbus_walk_resources, NULL);
974
975 if (ACPI_FAILURE(result))
976 goto acpi_walk_err;
90eedf0c
GH
977 if (hyperv_mmio.start && hyperv_mmio.end)
978 request_resource(&iomem_resource, &hyperv_mmio);
b0069f43 979 }
90f34535
S
980 ret_val = 0;
981
982acpi_walk_err:
b0069f43 983 complete(&probe_event);
90f34535 984 return ret_val;
b0069f43
S
985}
986
987static const struct acpi_device_id vmbus_acpi_device_ids[] = {
988 {"VMBUS", 0},
9d7b18d1 989 {"VMBus", 0},
b0069f43
S
990 {"", 0},
991};
992MODULE_DEVICE_TABLE(acpi, vmbus_acpi_device_ids);
993
994static struct acpi_driver vmbus_acpi_driver = {
995 .name = "vmbus",
996 .ids = vmbus_acpi_device_ids,
997 .ops = {
998 .add = vmbus_acpi_add,
999 },
1000};
1001
607c1a11 1002static int __init hv_acpi_init(void)
1168ac22 1003{
2dda95f8 1004 int ret, t;
b0069f43 1005
1f94ea81 1006 if (x86_hyper != &x86_hyper_ms_hyperv)
0592969e
JW
1007 return -ENODEV;
1008
b0069f43
S
1009 init_completion(&probe_event);
1010
1011 /*
1012 * Get irq resources first.
1013 */
0246604c
S
1014 ret = acpi_bus_register_driver(&vmbus_acpi_driver);
1015
b0069f43
S
1016 if (ret)
1017 return ret;
1018
2dda95f8
S
1019 t = wait_for_completion_timeout(&probe_event, 5*HZ);
1020 if (t == 0) {
1021 ret = -ETIMEDOUT;
1022 goto cleanup;
1023 }
b0069f43
S
1024
1025 if (irq <= 0) {
2dda95f8
S
1026 ret = -ENODEV;
1027 goto cleanup;
b0069f43
S
1028 }
1029
91fd799e
S
1030 ret = vmbus_bus_init(irq);
1031 if (ret)
2dda95f8
S
1032 goto cleanup;
1033
1034 return 0;
1035
1036cleanup:
1037 acpi_bus_unregister_driver(&vmbus_acpi_driver);
cf6a2eac 1038 hv_acpi_dev = NULL;
91fd799e 1039 return ret;
1168ac22
S
1040}
1041
93e5bd06
S
1042static void __exit vmbus_exit(void)
1043{
e72e7ac5
VK
1044 int cpu;
1045
09a19628 1046 vmbus_connection.conn_state = DISCONNECTED;
e086748c 1047 hv_synic_clockevents_cleanup();
76d388cd 1048 hv_remove_vmbus_irq();
93e5bd06
S
1049 vmbus_free_channels();
1050 bus_unregister(&hv_bus);
1051 hv_cleanup();
e72e7ac5
VK
1052 for_each_online_cpu(cpu)
1053 smp_call_function_single(cpu, hv_synic_cleanup, NULL, 1);
93e5bd06 1054 acpi_bus_unregister_driver(&vmbus_acpi_driver);
e513229b 1055 hv_cpu_hotplug_quirk(false);
09a19628 1056 vmbus_disconnect();
93e5bd06
S
1057}
1058
1168ac22 1059
90c9960e 1060MODULE_LICENSE("GPL");
3e7ee490 1061
43d4e119 1062subsys_initcall(hv_acpi_init);
93e5bd06 1063module_exit(vmbus_exit);