Commit | Line | Data |
---|---|---|
3e7ee490 | 1 | /* |
3e7ee490 HJ |
2 | * Copyright (c) 2009, Microsoft Corporation. |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify it | |
5 | * under the terms and conditions of the GNU General Public License, | |
6 | * version 2, as published by the Free Software Foundation. | |
7 | * | |
8 | * This program is distributed in the hope it will be useful, but WITHOUT | |
9 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
10 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | |
11 | * more details. | |
12 | * | |
13 | * You should have received a copy of the GNU General Public License along with | |
14 | * this program; if not, write to the Free Software Foundation, Inc., 59 Temple | |
15 | * Place - Suite 330, Boston, MA 02111-1307 USA. | |
16 | * | |
17 | * Authors: | |
18 | * Haiyang Zhang <haiyangz@microsoft.com> | |
19 | * Hank Janssen <hjanssen@microsoft.com> | |
3e7ee490 | 20 | */ |
0a46618d HJ |
21 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
22 | ||
5654e932 | 23 | #include <linux/kernel.h> |
0c3b7b2f S |
24 | #include <linux/sched.h> |
25 | #include <linux/wait.h> | |
a0086dc5 | 26 | #include <linux/mm.h> |
5a0e3ad6 | 27 | #include <linux/slab.h> |
c88c4e4c | 28 | #include <linux/module.h> |
46a97191 | 29 | #include <linux/hyperv.h> |
011a7c3c | 30 | #include <linux/uio.h> |
63d55b2a | 31 | #include <linux/interrupt.h> |
3f335ea2 | 32 | |
0f2a6619 | 33 | #include "hyperv_vmbus.h" |
3e7ee490 | 34 | |
e3fe0bb6 S |
35 | #define NUM_PAGES_SPANNED(addr, len) \ |
36 | ((PAGE_ALIGN(addr + len) >> PAGE_SHIFT) - (addr >> PAGE_SHIFT)) | |
37 | ||
3e189519 | 38 | /* |
fff41b2e | 39 | * vmbus_setevent- Trigger an event notification on the specified |
3e189519 | 40 | * channel. |
f4266e34 | 41 | */ |
fff41b2e | 42 | static void vmbus_setevent(struct vmbus_channel *channel) |
3e7ee490 | 43 | { |
39d70a4a | 44 | struct hv_monitor_page *monitorpage; |
3e7ee490 | 45 | |
3724287c S |
46 | /* |
47 | * For channels marked as in "low latency" mode | |
48 | * bypass the monitor page mechanism. | |
49 | */ | |
50 | if ((channel->offermsg.monitor_allocated) && | |
51 | (!channel->low_latency)) { | |
454f18a9 | 52 | /* Each u32 represents 32 channels */ |
22356585 | 53 | sync_set_bit(channel->offermsg.child_relid & 31, |
da9fcb72 | 54 | (unsigned long *) vmbus_connection.send_int_page + |
c50f7fb2 | 55 | (channel->offermsg.child_relid >> 5)); |
3e7ee490 | 56 | |
8681db44 GKH |
57 | /* Get the child to parent monitor page */ |
58 | monitorpage = vmbus_connection.monitor_pages[1]; | |
3e7ee490 | 59 | |
22356585 | 60 | sync_set_bit(channel->monitor_bit, |
f6feebe0 HZ |
61 | (unsigned long *)&monitorpage->trigger_group |
62 | [channel->monitor_grp].pending); | |
7c369f40 | 63 | |
f4266e34 | 64 | } else { |
21c3bef5 | 65 | vmbus_set_event(channel); |
3e7ee490 | 66 | } |
3e7ee490 HJ |
67 | } |
68 | ||
3e189519 | 69 | /* |
fff41b2e | 70 | * vmbus_open - Open the specified channel. |
f4266e34 | 71 | */ |
fff41b2e | 72 | int vmbus_open(struct vmbus_channel *newchannel, u32 send_ringbuffer_size, |
39d70a4a HZ |
73 | u32 recv_ringbuffer_size, void *userdata, u32 userdatalen, |
74 | void (*onchannelcallback)(void *context), void *context) | |
3e7ee490 | 75 | { |
0987ff69 | 76 | struct vmbus_channel_open_channel *open_msg; |
176fb9e3 | 77 | struct vmbus_channel_msginfo *open_info = NULL; |
dd0813b6 | 78 | unsigned long flags; |
08a9513f | 79 | int ret, err = 0; |
294409d2 | 80 | struct page *page; |
3e7ee490 | 81 | |
98f531b1 VK |
82 | if (send_ringbuffer_size % PAGE_SIZE || |
83 | recv_ringbuffer_size % PAGE_SIZE) | |
84 | return -EINVAL; | |
85 | ||
67fae053 | 86 | spin_lock_irqsave(&newchannel->lock, flags); |
e68d2971 S |
87 | if (newchannel->state == CHANNEL_OPEN_STATE) { |
88 | newchannel->state = CHANNEL_OPENING_STATE; | |
89 | } else { | |
67fae053 | 90 | spin_unlock_irqrestore(&newchannel->lock, flags); |
e68d2971 S |
91 | return -EINVAL; |
92 | } | |
67fae053 | 93 | spin_unlock_irqrestore(&newchannel->lock, flags); |
e68d2971 | 94 | |
c50f7fb2 HZ |
95 | newchannel->onchannel_callback = onchannelcallback; |
96 | newchannel->channel_callback_context = context; | |
3e7ee490 | 97 | |
454f18a9 | 98 | /* Allocate the ring buffer */ |
294409d2 S |
99 | page = alloc_pages_node(cpu_to_node(newchannel->target_cpu), |
100 | GFP_KERNEL|__GFP_ZERO, | |
101 | get_order(send_ringbuffer_size + | |
102 | recv_ringbuffer_size)); | |
103 | ||
104 | if (!page) | |
98f531b1 VK |
105 | page = alloc_pages(GFP_KERNEL|__GFP_ZERO, |
106 | get_order(send_ringbuffer_size + | |
107 | recv_ringbuffer_size)); | |
df3493e0 | 108 | |
98f531b1 | 109 | if (!page) { |
ac0d12b7 | 110 | err = -ENOMEM; |
98f531b1 | 111 | goto error_set_chnstate; |
ac0d12b7 | 112 | } |
3e7ee490 | 113 | |
9988ce68 | 114 | newchannel->ringbuffer_pages = page_address(page); |
c50f7fb2 | 115 | newchannel->ringbuffer_pagecount = (send_ringbuffer_size + |
39d70a4a | 116 | recv_ringbuffer_size) >> PAGE_SHIFT; |
3e7ee490 | 117 | |
9988ce68 VK |
118 | ret = hv_ringbuffer_init(&newchannel->outbound, page, |
119 | send_ringbuffer_size >> PAGE_SHIFT); | |
72a95cbc | 120 | |
fd4dc88e | 121 | if (ret != 0) { |
3324fb40 | 122 | err = ret; |
98f531b1 | 123 | goto error_free_pages; |
3324fb40 BP |
124 | } |
125 | ||
9988ce68 VK |
126 | ret = hv_ringbuffer_init(&newchannel->inbound, |
127 | &page[send_ringbuffer_size >> PAGE_SHIFT], | |
128 | recv_ringbuffer_size >> PAGE_SHIFT); | |
fd4dc88e | 129 | if (ret != 0) { |
3324fb40 | 130 | err = ret; |
98f531b1 | 131 | goto error_free_pages; |
3324fb40 | 132 | } |
3e7ee490 | 133 | |
3e7ee490 | 134 | |
454f18a9 | 135 | /* Establish the gpadl for the ring buffer */ |
c50f7fb2 | 136 | newchannel->ringbuffer_gpadlhandle = 0; |
3e7ee490 | 137 | |
fff41b2e | 138 | ret = vmbus_establish_gpadl(newchannel, |
9988ce68 VK |
139 | page_address(page), |
140 | send_ringbuffer_size + | |
141 | recv_ringbuffer_size, | |
142 | &newchannel->ringbuffer_gpadlhandle); | |
b94ef345 | 143 | |
fd4dc88e | 144 | if (ret != 0) { |
b94ef345 | 145 | err = ret; |
98f531b1 | 146 | goto error_free_pages; |
b94ef345 | 147 | } |
f4266e34 | 148 | |
454f18a9 | 149 | /* Create and init the channel open message */ |
176fb9e3 | 150 | open_info = kmalloc(sizeof(*open_info) + |
f4266e34 GKH |
151 | sizeof(struct vmbus_channel_open_channel), |
152 | GFP_KERNEL); | |
176fb9e3 | 153 | if (!open_info) { |
c3bf2e26 | 154 | err = -ENOMEM; |
98f531b1 | 155 | goto error_free_gpadl; |
c3bf2e26 | 156 | } |
3e7ee490 | 157 | |
176fb9e3 | 158 | init_completion(&open_info->waitevent); |
3e7ee490 | 159 | |
176fb9e3 | 160 | open_msg = (struct vmbus_channel_open_channel *)open_info->msg; |
0987ff69 S |
161 | open_msg->header.msgtype = CHANNELMSG_OPENCHANNEL; |
162 | open_msg->openid = newchannel->offermsg.child_relid; | |
163 | open_msg->child_relid = newchannel->offermsg.child_relid; | |
164 | open_msg->ringbuffer_gpadlhandle = newchannel->ringbuffer_gpadlhandle; | |
165 | open_msg->downstream_ringbuffer_pageoffset = send_ringbuffer_size >> | |
f4266e34 | 166 | PAGE_SHIFT; |
abbf3b2a | 167 | open_msg->target_vp = newchannel->target_vp; |
3e7ee490 | 168 | |
39d70a4a | 169 | if (userdatalen > MAX_USER_DEFINED_BYTES) { |
c827f944 | 170 | err = -EINVAL; |
98f531b1 | 171 | goto error_free_gpadl; |
c827f944 BP |
172 | } |
173 | ||
39d70a4a | 174 | if (userdatalen) |
0987ff69 | 175 | memcpy(open_msg->userdata, userdata, userdatalen); |
3e7ee490 | 176 | |
15b2f647 | 177 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
176fb9e3 | 178 | list_add_tail(&open_info->msglistentry, |
da9fcb72 | 179 | &vmbus_connection.chn_msg_list); |
15b2f647 | 180 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
3e7ee490 | 181 | |
0987ff69 | 182 | ret = vmbus_post_msg(open_msg, |
f4266e34 | 183 | sizeof(struct vmbus_channel_open_channel)); |
98e08702 | 184 | |
45d727ce S |
185 | if (ret != 0) { |
186 | err = ret; | |
98f531b1 | 187 | goto error_clean_msglist; |
45d727ce | 188 | } |
3e7ee490 | 189 | |
396e287f | 190 | wait_for_completion(&open_info->waitevent); |
0c3b7b2f | 191 | |
15b2f647 | 192 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
176fb9e3 | 193 | list_del(&open_info->msglistentry); |
15b2f647 | 194 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
3e7ee490 | 195 | |
ffc151f3 VK |
196 | if (open_info->response.open_result.status) { |
197 | err = -EAGAIN; | |
98f531b1 | 198 | goto error_free_gpadl; |
ffc151f3 | 199 | } |
e68d2971 | 200 | |
ffc151f3 | 201 | newchannel->state = CHANNEL_OPENED_STATE; |
176fb9e3 | 202 | kfree(open_info); |
ffc151f3 | 203 | return 0; |
c3bf2e26 | 204 | |
98f531b1 | 205 | error_clean_msglist: |
13925502 S |
206 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
207 | list_del(&open_info->msglistentry); | |
208 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); | |
209 | ||
98f531b1 | 210 | error_free_gpadl: |
40384e4b | 211 | vmbus_teardown_gpadl(newchannel, newchannel->ringbuffer_gpadlhandle); |
98f531b1 VK |
212 | kfree(open_info); |
213 | error_free_pages: | |
9988ce68 VK |
214 | hv_ringbuffer_cleanup(&newchannel->outbound); |
215 | hv_ringbuffer_cleanup(&newchannel->inbound); | |
216 | __free_pages(page, | |
217 | get_order(send_ringbuffer_size + recv_ringbuffer_size)); | |
98f531b1 | 218 | error_set_chnstate: |
ac0d12b7 | 219 | newchannel->state = CHANNEL_OPEN_STATE; |
c3bf2e26 | 220 | return err; |
3e7ee490 | 221 | } |
36ceadfc | 222 | EXPORT_SYMBOL_GPL(vmbus_open); |
3e7ee490 | 223 | |
5c23a1a5 DC |
224 | /* Used for Hyper-V Socket: a guest client's connect() to the host */ |
225 | int vmbus_send_tl_connect_request(const uuid_le *shv_guest_servie_id, | |
226 | const uuid_le *shv_host_servie_id) | |
227 | { | |
228 | struct vmbus_channel_tl_connect_request conn_msg; | |
229 | ||
230 | memset(&conn_msg, 0, sizeof(conn_msg)); | |
231 | conn_msg.header.msgtype = CHANNELMSG_TL_CONNECT_REQUEST; | |
232 | conn_msg.guest_endpoint_id = *shv_guest_servie_id; | |
233 | conn_msg.host_service_id = *shv_host_servie_id; | |
234 | ||
235 | return vmbus_post_msg(&conn_msg, sizeof(conn_msg)); | |
236 | } | |
237 | EXPORT_SYMBOL_GPL(vmbus_send_tl_connect_request); | |
238 | ||
3e189519 | 239 | /* |
fff41b2e | 240 | * create_gpadl_header - Creates a gpadl for the specified buffer |
f4266e34 | 241 | */ |
fff41b2e | 242 | static int create_gpadl_header(void *kbuffer, u32 size, |
4d637632 | 243 | struct vmbus_channel_msginfo **msginfo) |
3e7ee490 HJ |
244 | { |
245 | int i; | |
39d70a4a | 246 | int pagecount; |
39d70a4a HZ |
247 | struct vmbus_channel_gpadl_header *gpadl_header; |
248 | struct vmbus_channel_gpadl_body *gpadl_body; | |
249 | struct vmbus_channel_msginfo *msgheader; | |
250 | struct vmbus_channel_msginfo *msgbody = NULL; | |
251 | u32 msgsize; | |
3e7ee490 | 252 | |
39d70a4a | 253 | int pfnsum, pfncount, pfnleft, pfncurr, pfnsize; |
3e7ee490 | 254 | |
39d70a4a | 255 | pagecount = size >> PAGE_SHIFT; |
3e7ee490 | 256 | |
454f18a9 | 257 | /* do we need a gpadl body msg */ |
39d70a4a | 258 | pfnsize = MAX_SIZE_CHANNEL_MESSAGE - |
f4266e34 GKH |
259 | sizeof(struct vmbus_channel_gpadl_header) - |
260 | sizeof(struct gpa_range); | |
39d70a4a | 261 | pfncount = pfnsize / sizeof(u64); |
3e7ee490 | 262 | |
39d70a4a | 263 | if (pagecount > pfncount) { |
f4266e34 | 264 | /* we need a gpadl body */ |
454f18a9 | 265 | /* fill in the header */ |
39d70a4a | 266 | msgsize = sizeof(struct vmbus_channel_msginfo) + |
f4266e34 | 267 | sizeof(struct vmbus_channel_gpadl_header) + |
39d70a4a HZ |
268 | sizeof(struct gpa_range) + pfncount * sizeof(u64); |
269 | msgheader = kzalloc(msgsize, GFP_KERNEL); | |
270 | if (!msgheader) | |
d1c250bb | 271 | goto nomem; |
3e7ee490 | 272 | |
c50f7fb2 HZ |
273 | INIT_LIST_HEAD(&msgheader->submsglist); |
274 | msgheader->msgsize = msgsize; | |
3e7ee490 | 275 | |
39d70a4a | 276 | gpadl_header = (struct vmbus_channel_gpadl_header *) |
c50f7fb2 HZ |
277 | msgheader->msg; |
278 | gpadl_header->rangecount = 1; | |
279 | gpadl_header->range_buflen = sizeof(struct gpa_range) + | |
39d70a4a | 280 | pagecount * sizeof(u64); |
415f2287 HZ |
281 | gpadl_header->range[0].byte_offset = 0; |
282 | gpadl_header->range[0].byte_count = size; | |
39d70a4a | 283 | for (i = 0; i < pfncount; i++) |
b679ef73 HZ |
284 | gpadl_header->range[0].pfn_array[i] = slow_virt_to_phys( |
285 | kbuffer + PAGE_SIZE * i) >> PAGE_SHIFT; | |
39d70a4a | 286 | *msginfo = msgheader; |
3e7ee490 | 287 | |
39d70a4a HZ |
288 | pfnsum = pfncount; |
289 | pfnleft = pagecount - pfncount; | |
3e7ee490 | 290 | |
454f18a9 | 291 | /* how many pfns can we fit */ |
39d70a4a | 292 | pfnsize = MAX_SIZE_CHANNEL_MESSAGE - |
f4266e34 | 293 | sizeof(struct vmbus_channel_gpadl_body); |
39d70a4a | 294 | pfncount = pfnsize / sizeof(u64); |
3e7ee490 | 295 | |
454f18a9 | 296 | /* fill in the body */ |
39d70a4a HZ |
297 | while (pfnleft) { |
298 | if (pfnleft > pfncount) | |
299 | pfncurr = pfncount; | |
3e7ee490 | 300 | else |
39d70a4a | 301 | pfncurr = pfnleft; |
3e7ee490 | 302 | |
39d70a4a | 303 | msgsize = sizeof(struct vmbus_channel_msginfo) + |
f4266e34 | 304 | sizeof(struct vmbus_channel_gpadl_body) + |
39d70a4a HZ |
305 | pfncurr * sizeof(u64); |
306 | msgbody = kzalloc(msgsize, GFP_KERNEL); | |
f38cf9cc S |
307 | |
308 | if (!msgbody) { | |
309 | struct vmbus_channel_msginfo *pos = NULL; | |
310 | struct vmbus_channel_msginfo *tmp = NULL; | |
311 | /* | |
312 | * Free up all the allocated messages. | |
313 | */ | |
314 | list_for_each_entry_safe(pos, tmp, | |
315 | &msgheader->submsglist, | |
316 | msglistentry) { | |
317 | ||
318 | list_del(&pos->msglistentry); | |
319 | kfree(pos); | |
320 | } | |
321 | ||
d1c250bb | 322 | goto nomem; |
f38cf9cc S |
323 | } |
324 | ||
c50f7fb2 | 325 | msgbody->msgsize = msgsize; |
39d70a4a | 326 | gpadl_body = |
c50f7fb2 | 327 | (struct vmbus_channel_gpadl_body *)msgbody->msg; |
f4266e34 GKH |
328 | |
329 | /* | |
f4266e34 GKH |
330 | * Gpadl is u32 and we are using a pointer which could |
331 | * be 64-bit | |
f27df643 S |
332 | * This is governed by the guest/host protocol and |
333 | * so the hypervisor gurantees that this is ok. | |
f4266e34 | 334 | */ |
39d70a4a | 335 | for (i = 0; i < pfncurr; i++) |
b679ef73 HZ |
336 | gpadl_body->pfn[i] = slow_virt_to_phys( |
337 | kbuffer + PAGE_SIZE * (pfnsum + i)) >> | |
338 | PAGE_SHIFT; | |
3e7ee490 | 339 | |
454f18a9 | 340 | /* add to msg header */ |
c50f7fb2 HZ |
341 | list_add_tail(&msgbody->msglistentry, |
342 | &msgheader->submsglist); | |
39d70a4a HZ |
343 | pfnsum += pfncurr; |
344 | pfnleft -= pfncurr; | |
3e7ee490 | 345 | } |
f4266e34 | 346 | } else { |
454f18a9 | 347 | /* everything fits in a header */ |
39d70a4a | 348 | msgsize = sizeof(struct vmbus_channel_msginfo) + |
f4266e34 | 349 | sizeof(struct vmbus_channel_gpadl_header) + |
39d70a4a HZ |
350 | sizeof(struct gpa_range) + pagecount * sizeof(u64); |
351 | msgheader = kzalloc(msgsize, GFP_KERNEL); | |
352 | if (msgheader == NULL) | |
e3eb7cdd | 353 | goto nomem; |
4d637632 VK |
354 | |
355 | INIT_LIST_HEAD(&msgheader->submsglist); | |
c50f7fb2 | 356 | msgheader->msgsize = msgsize; |
39d70a4a HZ |
357 | |
358 | gpadl_header = (struct vmbus_channel_gpadl_header *) | |
c50f7fb2 HZ |
359 | msgheader->msg; |
360 | gpadl_header->rangecount = 1; | |
361 | gpadl_header->range_buflen = sizeof(struct gpa_range) + | |
39d70a4a | 362 | pagecount * sizeof(u64); |
415f2287 HZ |
363 | gpadl_header->range[0].byte_offset = 0; |
364 | gpadl_header->range[0].byte_count = size; | |
39d70a4a | 365 | for (i = 0; i < pagecount; i++) |
b679ef73 HZ |
366 | gpadl_header->range[0].pfn_array[i] = slow_virt_to_phys( |
367 | kbuffer + PAGE_SIZE * i) >> PAGE_SHIFT; | |
39d70a4a HZ |
368 | |
369 | *msginfo = msgheader; | |
3e7ee490 HJ |
370 | } |
371 | ||
372 | return 0; | |
d1c250bb | 373 | nomem: |
39d70a4a HZ |
374 | kfree(msgheader); |
375 | kfree(msgbody); | |
d1c250bb | 376 | return -ENOMEM; |
3e7ee490 HJ |
377 | } |
378 | ||
3e189519 | 379 | /* |
fff41b2e | 380 | * vmbus_establish_gpadl - Estabish a GPADL for the specified buffer |
f4266e34 | 381 | * |
39d70a4a | 382 | * @channel: a channel |
b679ef73 | 383 | * @kbuffer: from kmalloc or vmalloc |
39d70a4a HZ |
384 | * @size: page-size multiple |
385 | * @gpadl_handle: some funky thing | |
f4266e34 | 386 | */ |
fff41b2e | 387 | int vmbus_establish_gpadl(struct vmbus_channel *channel, void *kbuffer, |
39d70a4a | 388 | u32 size, u32 *gpadl_handle) |
3e7ee490 | 389 | { |
39d70a4a HZ |
390 | struct vmbus_channel_gpadl_header *gpadlmsg; |
391 | struct vmbus_channel_gpadl_body *gpadl_body; | |
39d70a4a | 392 | struct vmbus_channel_msginfo *msginfo = NULL; |
7cc80c98 | 393 | struct vmbus_channel_msginfo *submsginfo, *tmp; |
53af545b | 394 | struct list_head *curr; |
39d70a4a | 395 | u32 next_gpadl_handle; |
dd0813b6 | 396 | unsigned long flags; |
c3bf2e26 | 397 | int ret = 0; |
3e7ee490 | 398 | |
9f52a163 S |
399 | next_gpadl_handle = |
400 | (atomic_inc_return(&vmbus_connection.next_gpadl_handle) - 1); | |
3e7ee490 | 401 | |
4d637632 | 402 | ret = create_gpadl_header(kbuffer, size, &msginfo); |
c3bf2e26 BP |
403 | if (ret) |
404 | return ret; | |
3e7ee490 | 405 | |
9568a193 | 406 | init_completion(&msginfo->waitevent); |
c3bf2e26 | 407 | |
c50f7fb2 HZ |
408 | gpadlmsg = (struct vmbus_channel_gpadl_header *)msginfo->msg; |
409 | gpadlmsg->header.msgtype = CHANNELMSG_GPADL_HEADER; | |
410 | gpadlmsg->child_relid = channel->offermsg.child_relid; | |
411 | gpadlmsg->gpadl = next_gpadl_handle; | |
3e7ee490 | 412 | |
3e7ee490 | 413 | |
15b2f647 | 414 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
c50f7fb2 | 415 | list_add_tail(&msginfo->msglistentry, |
da9fcb72 | 416 | &vmbus_connection.chn_msg_list); |
3e7ee490 | 417 | |
15b2f647 | 418 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
3e7ee490 | 419 | |
c6977677 | 420 | ret = vmbus_post_msg(gpadlmsg, msginfo->msgsize - |
39d70a4a | 421 | sizeof(*msginfo)); |
98e08702 | 422 | if (ret != 0) |
00d760b0 | 423 | goto cleanup; |
3e7ee490 | 424 | |
4d637632 VK |
425 | list_for_each(curr, &msginfo->submsglist) { |
426 | submsginfo = (struct vmbus_channel_msginfo *)curr; | |
427 | gpadl_body = | |
428 | (struct vmbus_channel_gpadl_body *)submsginfo->msg; | |
53af545b | 429 | |
4d637632 VK |
430 | gpadl_body->header.msgtype = |
431 | CHANNELMSG_GPADL_BODY; | |
432 | gpadl_body->gpadl = next_gpadl_handle; | |
3e7ee490 | 433 | |
4d637632 VK |
434 | ret = vmbus_post_msg(gpadl_body, |
435 | submsginfo->msgsize - | |
436 | sizeof(*submsginfo)); | |
437 | if (ret != 0) | |
438 | goto cleanup; | |
3e7ee490 | 439 | |
3e7ee490 | 440 | } |
72c6b71c | 441 | wait_for_completion(&msginfo->waitevent); |
3e7ee490 | 442 | |
454f18a9 | 443 | /* At this point, we received the gpadl created msg */ |
c50f7fb2 | 444 | *gpadl_handle = gpadlmsg->gpadl; |
3e7ee490 | 445 | |
00d760b0 | 446 | cleanup: |
15b2f647 | 447 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
c50f7fb2 | 448 | list_del(&msginfo->msglistentry); |
15b2f647 | 449 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
7cc80c98 VK |
450 | list_for_each_entry_safe(submsginfo, tmp, &msginfo->submsglist, |
451 | msglistentry) { | |
452 | kfree(submsginfo); | |
453 | } | |
3e7ee490 | 454 | |
39d70a4a | 455 | kfree(msginfo); |
3e7ee490 HJ |
456 | return ret; |
457 | } | |
98873724 | 458 | EXPORT_SYMBOL_GPL(vmbus_establish_gpadl); |
3e7ee490 | 459 | |
3e189519 | 460 | /* |
fff41b2e | 461 | * vmbus_teardown_gpadl -Teardown the specified GPADL handle |
f4266e34 | 462 | */ |
fff41b2e | 463 | int vmbus_teardown_gpadl(struct vmbus_channel *channel, u32 gpadl_handle) |
3e7ee490 | 464 | { |
82250213 | 465 | struct vmbus_channel_gpadl_teardown *msg; |
aded7165 | 466 | struct vmbus_channel_msginfo *info; |
dd0813b6 | 467 | unsigned long flags; |
66be6530 | 468 | int ret; |
3e7ee490 | 469 | |
f4266e34 GKH |
470 | info = kmalloc(sizeof(*info) + |
471 | sizeof(struct vmbus_channel_gpadl_teardown), GFP_KERNEL); | |
c3bf2e26 BP |
472 | if (!info) |
473 | return -ENOMEM; | |
3e7ee490 | 474 | |
9568a193 | 475 | init_completion(&info->waitevent); |
3e7ee490 | 476 | |
c50f7fb2 | 477 | msg = (struct vmbus_channel_gpadl_teardown *)info->msg; |
3e7ee490 | 478 | |
c50f7fb2 HZ |
479 | msg->header.msgtype = CHANNELMSG_GPADL_TEARDOWN; |
480 | msg->child_relid = channel->offermsg.child_relid; | |
481 | msg->gpadl = gpadl_handle; | |
3e7ee490 | 482 | |
15b2f647 | 483 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
c50f7fb2 | 484 | list_add_tail(&info->msglistentry, |
da9fcb72 | 485 | &vmbus_connection.chn_msg_list); |
15b2f647 | 486 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
c6977677 | 487 | ret = vmbus_post_msg(msg, |
f4266e34 | 488 | sizeof(struct vmbus_channel_gpadl_teardown)); |
3e7ee490 | 489 | |
66be6530 S |
490 | if (ret) |
491 | goto post_msg_err; | |
492 | ||
493 | wait_for_completion(&info->waitevent); | |
3e7ee490 | 494 | |
66be6530 | 495 | post_msg_err: |
15b2f647 | 496 | spin_lock_irqsave(&vmbus_connection.channelmsg_lock, flags); |
c50f7fb2 | 497 | list_del(&info->msglistentry); |
15b2f647 | 498 | spin_unlock_irqrestore(&vmbus_connection.channelmsg_lock, flags); |
3e7ee490 | 499 | |
8c69f52a | 500 | kfree(info); |
3e7ee490 HJ |
501 | return ret; |
502 | } | |
18726d7a | 503 | EXPORT_SYMBOL_GPL(vmbus_teardown_gpadl); |
3e7ee490 | 504 | |
d3ba720d S |
505 | static void reset_channel_cb(void *arg) |
506 | { | |
507 | struct vmbus_channel *channel = arg; | |
508 | ||
509 | channel->onchannel_callback = NULL; | |
510 | } | |
511 | ||
98d731bb | 512 | static int vmbus_close_internal(struct vmbus_channel *channel) |
3e7ee490 | 513 | { |
82250213 | 514 | struct vmbus_channel_close_channel *msg; |
f4266e34 | 515 | int ret; |
3e7ee490 | 516 | |
63d55b2a DC |
517 | /* |
518 | * process_chn_event(), running in the tasklet, can race | |
519 | * with vmbus_close_internal() in the case of SMP guest, e.g., when | |
520 | * the former is accessing channel->inbound.ring_buffer, the latter | |
521 | * could be freeing the ring_buffer pages. | |
522 | * | |
523 | * To resolve the race, we can serialize them by disabling the | |
524 | * tasklet when the latter is running here. | |
525 | */ | |
638fea33 | 526 | hv_event_tasklet_disable(channel); |
63d55b2a | 527 | |
64b7faf9 DC |
528 | /* |
529 | * In case a device driver's probe() fails (e.g., | |
530 | * util_probe() -> vmbus_open() returns -ENOMEM) and the device is | |
531 | * rescinded later (e.g., we dynamically disble an Integrated Service | |
532 | * in Hyper-V Manager), the driver's remove() invokes vmbus_close(): | |
533 | * here we should skip most of the below cleanup work. | |
534 | */ | |
535 | if (channel->state != CHANNEL_OPENED_STATE) { | |
536 | ret = -EINVAL; | |
537 | goto out; | |
538 | } | |
539 | ||
e68d2971 S |
540 | channel->state = CHANNEL_OPEN_STATE; |
541 | channel->sc_creation_callback = NULL; | |
454f18a9 | 542 | /* Stop callback and cancel the timer asap */ |
2115b561 S |
543 | if (channel->target_cpu != get_cpu()) { |
544 | put_cpu(); | |
d3ba720d S |
545 | smp_call_function_single(channel->target_cpu, reset_channel_cb, |
546 | channel, true); | |
2115b561 | 547 | } else { |
d3ba720d | 548 | reset_channel_cb(channel); |
2115b561 S |
549 | put_cpu(); |
550 | } | |
3e7ee490 | 551 | |
454f18a9 | 552 | /* Send a closing message */ |
3e7ee490 | 553 | |
e9a27a9f | 554 | msg = &channel->close_msg.msg; |
3e7ee490 | 555 | |
c50f7fb2 HZ |
556 | msg->header.msgtype = CHANNELMSG_CLOSECHANNEL; |
557 | msg->child_relid = channel->offermsg.child_relid; | |
3e7ee490 | 558 | |
c6977677 | 559 | ret = vmbus_post_msg(msg, sizeof(struct vmbus_channel_close_channel)); |
3e7ee490 | 560 | |
98d731bb S |
561 | if (ret) { |
562 | pr_err("Close failed: close post msg return is %d\n", ret); | |
563 | /* | |
564 | * If we failed to post the close msg, | |
565 | * it is perhaps better to leak memory. | |
566 | */ | |
63d55b2a | 567 | goto out; |
98d731bb S |
568 | } |
569 | ||
454f18a9 | 570 | /* Tear down the gpadl for the channel's ring buffer */ |
98d731bb S |
571 | if (channel->ringbuffer_gpadlhandle) { |
572 | ret = vmbus_teardown_gpadl(channel, | |
573 | channel->ringbuffer_gpadlhandle); | |
574 | if (ret) { | |
575 | pr_err("Close failed: teardown gpadl return %d\n", ret); | |
576 | /* | |
577 | * If we failed to teardown gpadl, | |
578 | * it is perhaps better to leak memory. | |
579 | */ | |
63d55b2a | 580 | goto out; |
98d731bb S |
581 | } |
582 | } | |
3e7ee490 | 583 | |
454f18a9 | 584 | /* Cleanup the ring buffers for this channel */ |
2dba688b S |
585 | hv_ringbuffer_cleanup(&channel->outbound); |
586 | hv_ringbuffer_cleanup(&channel->inbound); | |
3e7ee490 | 587 | |
df3493e0 S |
588 | free_pages((unsigned long)channel->ringbuffer_pages, |
589 | get_order(channel->ringbuffer_pagecount * PAGE_SIZE)); | |
3e7ee490 | 590 | |
63d55b2a | 591 | out: |
638fea33 | 592 | hv_event_tasklet_enable(channel); |
63d55b2a | 593 | |
98d731bb | 594 | return ret; |
3e7ee490 | 595 | } |
e68d2971 S |
596 | |
597 | /* | |
598 | * vmbus_close - Close the specified channel | |
599 | */ | |
600 | void vmbus_close(struct vmbus_channel *channel) | |
601 | { | |
602 | struct list_head *cur, *tmp; | |
603 | struct vmbus_channel *cur_channel; | |
604 | ||
605 | if (channel->primary_channel != NULL) { | |
606 | /* | |
607 | * We will only close sub-channels when | |
608 | * the primary is closed. | |
609 | */ | |
610 | return; | |
611 | } | |
612 | /* | |
613 | * Close all the sub-channels first and then close the | |
614 | * primary channel. | |
615 | */ | |
616 | list_for_each_safe(cur, tmp, &channel->sc_list) { | |
617 | cur_channel = list_entry(cur, struct vmbus_channel, sc_list); | |
618 | if (cur_channel->state != CHANNEL_OPENED_STATE) | |
619 | continue; | |
620 | vmbus_close_internal(cur_channel); | |
621 | } | |
622 | /* | |
623 | * Now close the primary. | |
624 | */ | |
625 | vmbus_close_internal(channel); | |
626 | } | |
70bfa307 | 627 | EXPORT_SYMBOL_GPL(vmbus_close); |
3e7ee490 | 628 | |
e9395e3f | 629 | int vmbus_sendpacket_ctl(struct vmbus_channel *channel, void *buffer, |
39d70a4a | 630 | u32 bufferlen, u64 requestid, |
e9395e3f | 631 | enum vmbus_packet_type type, u32 flags, bool kick_q) |
3e7ee490 | 632 | { |
8dc0a06a | 633 | struct vmpacket_descriptor desc; |
39d70a4a | 634 | u32 packetlen = sizeof(struct vmpacket_descriptor) + bufferlen; |
73509681 | 635 | u32 packetlen_aligned = ALIGN(packetlen, sizeof(u64)); |
011a7c3c | 636 | struct kvec bufferlist[3]; |
39d70a4a | 637 | u64 aligned_data = 0; |
f4266e34 | 638 | int ret; |
98fa8cf4 | 639 | bool signal = false; |
fe760e4d | 640 | bool lock = channel->acquire_ring_lock; |
b81658cf | 641 | int num_vecs = ((bufferlen != 0) ? 3 : 1); |
3e7ee490 | 642 | |
3e7ee490 | 643 | |
454f18a9 | 644 | /* Setup the descriptor */ |
415f2287 HZ |
645 | desc.type = type; /* VmbusPacketTypeDataInBand; */ |
646 | desc.flags = flags; /* VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; */ | |
f4266e34 | 647 | /* in 8-bytes granularity */ |
415f2287 HZ |
648 | desc.offset8 = sizeof(struct vmpacket_descriptor) >> 3; |
649 | desc.len8 = (u16)(packetlen_aligned >> 3); | |
650 | desc.trans_id = requestid; | |
3e7ee490 | 651 | |
011a7c3c S |
652 | bufferlist[0].iov_base = &desc; |
653 | bufferlist[0].iov_len = sizeof(struct vmpacket_descriptor); | |
654 | bufferlist[1].iov_base = buffer; | |
655 | bufferlist[1].iov_len = bufferlen; | |
656 | bufferlist[2].iov_base = &aligned_data; | |
657 | bufferlist[2].iov_len = (packetlen_aligned - packetlen); | |
3e7ee490 | 658 | |
b81658cf | 659 | ret = hv_ringbuffer_write(&channel->outbound, bufferlist, num_vecs, |
ccef9bcc | 660 | &signal, lock, channel->signal_policy); |
3e7ee490 | 661 | |
5f5cc817 S |
662 | /* |
663 | * Signalling the host is conditional on many factors: | |
664 | * 1. The ring state changed from being empty to non-empty. | |
665 | * This is tracked by the variable "signal". | |
666 | * 2. The variable kick_q tracks if more data will be placed | |
667 | * on the ring. We will not signal if more data is | |
668 | * to be placed. | |
669 | * | |
8599846d S |
670 | * Based on the channel signal state, we will decide |
671 | * which signaling policy will be applied. | |
672 | * | |
5f5cc817 S |
673 | * If we cannot write to the ring-buffer; signal the host |
674 | * even if we may not have written anything. This is a rare | |
675 | * enough condition that it should not matter. | |
5f363bc3 DC |
676 | * NOTE: in this case, the hvsock channel is an exception, because |
677 | * it looks the host side's hvsock implementation has a throttling | |
678 | * mechanism which can hurt the performance otherwise. | |
5f5cc817 | 679 | */ |
8599846d | 680 | |
5f363bc3 DC |
681 | if (((ret == 0) && kick_q && signal) || |
682 | (ret && !is_hvsock_channel(channel))) | |
fff41b2e | 683 | vmbus_setevent(channel); |
3e7ee490 | 684 | |
3e7ee490 HJ |
685 | return ret; |
686 | } | |
e9395e3f S |
687 | EXPORT_SYMBOL(vmbus_sendpacket_ctl); |
688 | ||
689 | /** | |
690 | * vmbus_sendpacket() - Send the specified buffer on the given channel | |
691 | * @channel: Pointer to vmbus_channel structure. | |
692 | * @buffer: Pointer to the buffer you want to receive the data into. | |
693 | * @bufferlen: Maximum size of what the the buffer will hold | |
694 | * @requestid: Identifier of the request | |
695 | * @type: Type of packet that is being send e.g. negotiate, time | |
696 | * packet etc. | |
697 | * | |
698 | * Sends data in @buffer directly to hyper-v via the vmbus | |
699 | * This will send the data unparsed to hyper-v. | |
700 | * | |
701 | * Mainly used by Hyper-V drivers. | |
702 | */ | |
703 | int vmbus_sendpacket(struct vmbus_channel *channel, void *buffer, | |
704 | u32 bufferlen, u64 requestid, | |
705 | enum vmbus_packet_type type, u32 flags) | |
706 | { | |
707 | return vmbus_sendpacket_ctl(channel, buffer, bufferlen, requestid, | |
708 | type, flags, true); | |
709 | } | |
fff41b2e | 710 | EXPORT_SYMBOL(vmbus_sendpacket); |
3e7ee490 | 711 | |
3e189519 | 712 | /* |
87e93d61 S |
713 | * vmbus_sendpacket_pagebuffer_ctl - Send a range of single-page buffer |
714 | * packets using a GPADL Direct packet type. This interface allows you | |
715 | * to control notifying the host. This will be useful for sending | |
716 | * batched data. Also the sender can control the send flags | |
717 | * explicitly. | |
f4266e34 | 718 | */ |
87e93d61 | 719 | int vmbus_sendpacket_pagebuffer_ctl(struct vmbus_channel *channel, |
39d70a4a HZ |
720 | struct hv_page_buffer pagebuffers[], |
721 | u32 pagecount, void *buffer, u32 bufferlen, | |
87e93d61 S |
722 | u64 requestid, |
723 | u32 flags, | |
724 | bool kick_q) | |
3e7ee490 | 725 | { |
f4266e34 GKH |
726 | int ret; |
727 | int i; | |
430a8e9a | 728 | struct vmbus_channel_packet_page_buffer desc; |
39d70a4a HZ |
729 | u32 descsize; |
730 | u32 packetlen; | |
731 | u32 packetlen_aligned; | |
011a7c3c | 732 | struct kvec bufferlist[3]; |
39d70a4a | 733 | u64 aligned_data = 0; |
98fa8cf4 | 734 | bool signal = false; |
fe760e4d | 735 | bool lock = channel->acquire_ring_lock; |
3e7ee490 | 736 | |
39d70a4a | 737 | if (pagecount > MAX_PAGE_BUFFER_COUNT) |
002b53ea | 738 | return -EINVAL; |
3e7ee490 | 739 | |
3e7ee490 | 740 | |
f4266e34 | 741 | /* |
430a8e9a | 742 | * Adjust the size down since vmbus_channel_packet_page_buffer is the |
f4266e34 GKH |
743 | * largest size we support |
744 | */ | |
39d70a4a HZ |
745 | descsize = sizeof(struct vmbus_channel_packet_page_buffer) - |
746 | ((MAX_PAGE_BUFFER_COUNT - pagecount) * | |
f4266e34 | 747 | sizeof(struct hv_page_buffer)); |
39d70a4a | 748 | packetlen = descsize + bufferlen; |
73509681 | 749 | packetlen_aligned = ALIGN(packetlen, sizeof(u64)); |
3e7ee490 | 750 | |
454f18a9 | 751 | /* Setup the descriptor */ |
415f2287 | 752 | desc.type = VM_PKT_DATA_USING_GPA_DIRECT; |
87e93d61 | 753 | desc.flags = flags; |
39d70a4a HZ |
754 | desc.dataoffset8 = descsize >> 3; /* in 8-bytes grandularity */ |
755 | desc.length8 = (u16)(packetlen_aligned >> 3); | |
756 | desc.transactionid = requestid; | |
757 | desc.rangecount = pagecount; | |
758 | ||
759 | for (i = 0; i < pagecount; i++) { | |
ca623ad3 HZ |
760 | desc.range[i].len = pagebuffers[i].len; |
761 | desc.range[i].offset = pagebuffers[i].offset; | |
762 | desc.range[i].pfn = pagebuffers[i].pfn; | |
3e7ee490 HJ |
763 | } |
764 | ||
011a7c3c S |
765 | bufferlist[0].iov_base = &desc; |
766 | bufferlist[0].iov_len = descsize; | |
767 | bufferlist[1].iov_base = buffer; | |
768 | bufferlist[1].iov_len = bufferlen; | |
769 | bufferlist[2].iov_base = &aligned_data; | |
770 | bufferlist[2].iov_len = (packetlen_aligned - packetlen); | |
3e7ee490 | 771 | |
fe760e4d | 772 | ret = hv_ringbuffer_write(&channel->outbound, bufferlist, 3, |
ccef9bcc | 773 | &signal, lock, channel->signal_policy); |
3e7ee490 | 774 | |
5f5cc817 S |
775 | /* |
776 | * Signalling the host is conditional on many factors: | |
777 | * 1. The ring state changed from being empty to non-empty. | |
778 | * This is tracked by the variable "signal". | |
779 | * 2. The variable kick_q tracks if more data will be placed | |
780 | * on the ring. We will not signal if more data is | |
781 | * to be placed. | |
782 | * | |
8599846d S |
783 | * Based on the channel signal state, we will decide |
784 | * which signaling policy will be applied. | |
785 | * | |
5f5cc817 S |
786 | * If we cannot write to the ring-buffer; signal the host |
787 | * even if we may not have written anything. This is a rare | |
788 | * enough condition that it should not matter. | |
789 | */ | |
8599846d | 790 | |
5f5cc817 | 791 | if (((ret == 0) && kick_q && signal) || (ret)) |
fff41b2e | 792 | vmbus_setevent(channel); |
3e7ee490 | 793 | |
3e7ee490 HJ |
794 | return ret; |
795 | } | |
b3a19b36 | 796 | EXPORT_SYMBOL_GPL(vmbus_sendpacket_pagebuffer_ctl); |
87e93d61 S |
797 | |
798 | /* | |
799 | * vmbus_sendpacket_pagebuffer - Send a range of single-page buffer | |
800 | * packets using a GPADL Direct packet type. | |
801 | */ | |
802 | int vmbus_sendpacket_pagebuffer(struct vmbus_channel *channel, | |
803 | struct hv_page_buffer pagebuffers[], | |
804 | u32 pagecount, void *buffer, u32 bufferlen, | |
805 | u64 requestid) | |
806 | { | |
807 | u32 flags = VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; | |
808 | return vmbus_sendpacket_pagebuffer_ctl(channel, pagebuffers, pagecount, | |
809 | buffer, bufferlen, requestid, | |
810 | flags, true); | |
811 | ||
812 | } | |
713efeb4 | 813 | EXPORT_SYMBOL_GPL(vmbus_sendpacket_pagebuffer); |
3e7ee490 | 814 | |
d61031ee S |
815 | /* |
816 | * vmbus_sendpacket_multipagebuffer - Send a multi-page buffer packet | |
817 | * using a GPADL Direct packet type. | |
818 | * The buffer includes the vmbus descriptor. | |
819 | */ | |
820 | int vmbus_sendpacket_mpb_desc(struct vmbus_channel *channel, | |
821 | struct vmbus_packet_mpb_array *desc, | |
822 | u32 desc_size, | |
823 | void *buffer, u32 bufferlen, u64 requestid) | |
824 | { | |
825 | int ret; | |
826 | u32 packetlen; | |
827 | u32 packetlen_aligned; | |
828 | struct kvec bufferlist[3]; | |
829 | u64 aligned_data = 0; | |
830 | bool signal = false; | |
fe760e4d | 831 | bool lock = channel->acquire_ring_lock; |
d61031ee S |
832 | |
833 | packetlen = desc_size + bufferlen; | |
834 | packetlen_aligned = ALIGN(packetlen, sizeof(u64)); | |
835 | ||
836 | /* Setup the descriptor */ | |
837 | desc->type = VM_PKT_DATA_USING_GPA_DIRECT; | |
838 | desc->flags = VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; | |
839 | desc->dataoffset8 = desc_size >> 3; /* in 8-bytes grandularity */ | |
840 | desc->length8 = (u16)(packetlen_aligned >> 3); | |
841 | desc->transactionid = requestid; | |
842 | desc->rangecount = 1; | |
843 | ||
844 | bufferlist[0].iov_base = desc; | |
845 | bufferlist[0].iov_len = desc_size; | |
846 | bufferlist[1].iov_base = buffer; | |
847 | bufferlist[1].iov_len = bufferlen; | |
848 | bufferlist[2].iov_base = &aligned_data; | |
849 | bufferlist[2].iov_len = (packetlen_aligned - packetlen); | |
850 | ||
fe760e4d | 851 | ret = hv_ringbuffer_write(&channel->outbound, bufferlist, 3, |
ccef9bcc | 852 | &signal, lock, channel->signal_policy); |
d61031ee S |
853 | |
854 | if (ret == 0 && signal) | |
855 | vmbus_setevent(channel); | |
856 | ||
857 | return ret; | |
858 | } | |
859 | EXPORT_SYMBOL_GPL(vmbus_sendpacket_mpb_desc); | |
860 | ||
3e189519 | 861 | /* |
fff41b2e | 862 | * vmbus_sendpacket_multipagebuffer - Send a multi-page buffer packet |
3e189519 | 863 | * using a GPADL Direct packet type. |
f4266e34 | 864 | */ |
fff41b2e | 865 | int vmbus_sendpacket_multipagebuffer(struct vmbus_channel *channel, |
39d70a4a HZ |
866 | struct hv_multipage_buffer *multi_pagebuffer, |
867 | void *buffer, u32 bufferlen, u64 requestid) | |
3e7ee490 | 868 | { |
f4266e34 | 869 | int ret; |
430a8e9a | 870 | struct vmbus_channel_packet_multipage_buffer desc; |
39d70a4a HZ |
871 | u32 descsize; |
872 | u32 packetlen; | |
873 | u32 packetlen_aligned; | |
011a7c3c | 874 | struct kvec bufferlist[3]; |
39d70a4a | 875 | u64 aligned_data = 0; |
98fa8cf4 | 876 | bool signal = false; |
fe760e4d | 877 | bool lock = channel->acquire_ring_lock; |
ca623ad3 HZ |
878 | u32 pfncount = NUM_PAGES_SPANNED(multi_pagebuffer->offset, |
879 | multi_pagebuffer->len); | |
3e7ee490 | 880 | |
24b8a406 | 881 | if (pfncount > MAX_MULTIPAGE_BUFFER_COUNT) |
002b53ea | 882 | return -EINVAL; |
3e7ee490 | 883 | |
f4266e34 | 884 | /* |
430a8e9a | 885 | * Adjust the size down since vmbus_channel_packet_multipage_buffer is |
f4266e34 GKH |
886 | * the largest size we support |
887 | */ | |
39d70a4a HZ |
888 | descsize = sizeof(struct vmbus_channel_packet_multipage_buffer) - |
889 | ((MAX_MULTIPAGE_BUFFER_COUNT - pfncount) * | |
f4266e34 | 890 | sizeof(u64)); |
39d70a4a | 891 | packetlen = descsize + bufferlen; |
73509681 | 892 | packetlen_aligned = ALIGN(packetlen, sizeof(u64)); |
3e7ee490 | 893 | |
3e7ee490 | 894 | |
454f18a9 | 895 | /* Setup the descriptor */ |
415f2287 | 896 | desc.type = VM_PKT_DATA_USING_GPA_DIRECT; |
430a8e9a | 897 | desc.flags = VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; |
39d70a4a HZ |
898 | desc.dataoffset8 = descsize >> 3; /* in 8-bytes grandularity */ |
899 | desc.length8 = (u16)(packetlen_aligned >> 3); | |
900 | desc.transactionid = requestid; | |
430a8e9a | 901 | desc.rangecount = 1; |
3e7ee490 | 902 | |
ca623ad3 HZ |
903 | desc.range.len = multi_pagebuffer->len; |
904 | desc.range.offset = multi_pagebuffer->offset; | |
3e7ee490 | 905 | |
ca623ad3 | 906 | memcpy(desc.range.pfn_array, multi_pagebuffer->pfn_array, |
39d70a4a | 907 | pfncount * sizeof(u64)); |
3e7ee490 | 908 | |
011a7c3c S |
909 | bufferlist[0].iov_base = &desc; |
910 | bufferlist[0].iov_len = descsize; | |
911 | bufferlist[1].iov_base = buffer; | |
912 | bufferlist[1].iov_len = bufferlen; | |
913 | bufferlist[2].iov_base = &aligned_data; | |
914 | bufferlist[2].iov_len = (packetlen_aligned - packetlen); | |
3e7ee490 | 915 | |
fe760e4d | 916 | ret = hv_ringbuffer_write(&channel->outbound, bufferlist, 3, |
ccef9bcc | 917 | &signal, lock, channel->signal_policy); |
3e7ee490 | 918 | |
98fa8cf4 | 919 | if (ret == 0 && signal) |
fff41b2e | 920 | vmbus_setevent(channel); |
3e7ee490 | 921 | |
3e7ee490 HJ |
922 | return ret; |
923 | } | |
4cb106fa | 924 | EXPORT_SYMBOL_GPL(vmbus_sendpacket_multipagebuffer); |
c88c4e4c HJ |
925 | |
926 | /** | |
fff41b2e | 927 | * vmbus_recvpacket() - Retrieve the user packet on the specified channel |
39d70a4a HZ |
928 | * @channel: Pointer to vmbus_channel structure. |
929 | * @buffer: Pointer to the buffer you want to receive the data into. | |
930 | * @bufferlen: Maximum size of what the the buffer will hold | |
931 | * @buffer_actual_len: The actual size of the data after it was received | |
932 | * @requestid: Identifier of the request | |
c88c4e4c HJ |
933 | * |
934 | * Receives directly from the hyper-v vmbus and puts the data it received | |
935 | * into Buffer. This will receive the data unparsed from hyper-v. | |
936 | * | |
937 | * Mainly used by Hyper-V drivers. | |
f4266e34 | 938 | */ |
667d3740 VK |
939 | static inline int |
940 | __vmbus_recvpacket(struct vmbus_channel *channel, void *buffer, | |
941 | u32 bufferlen, u32 *buffer_actual_len, u64 *requestid, | |
942 | bool raw) | |
3e7ee490 | 943 | { |
3e7ee490 | 944 | int ret; |
c2b8e520 | 945 | bool signal = false; |
3e7ee490 | 946 | |
940b68e2 VK |
947 | ret = hv_ringbuffer_read(&channel->inbound, buffer, bufferlen, |
948 | buffer_actual_len, requestid, &signal, raw); | |
3e7ee490 | 949 | |
c2b8e520 S |
950 | if (signal) |
951 | vmbus_setevent(channel); | |
3e7ee490 | 952 | |
667d3740 VK |
953 | return ret; |
954 | } | |
955 | ||
956 | int vmbus_recvpacket(struct vmbus_channel *channel, void *buffer, | |
957 | u32 bufferlen, u32 *buffer_actual_len, | |
958 | u64 *requestid) | |
959 | { | |
960 | return __vmbus_recvpacket(channel, buffer, bufferlen, | |
961 | buffer_actual_len, requestid, false); | |
3e7ee490 | 962 | } |
fff41b2e | 963 | EXPORT_SYMBOL(vmbus_recvpacket); |
3e7ee490 | 964 | |
3e189519 | 965 | /* |
fff41b2e | 966 | * vmbus_recvpacket_raw - Retrieve the raw packet on the specified channel |
f4266e34 | 967 | */ |
fff41b2e | 968 | int vmbus_recvpacket_raw(struct vmbus_channel *channel, void *buffer, |
39d70a4a HZ |
969 | u32 bufferlen, u32 *buffer_actual_len, |
970 | u64 *requestid) | |
3e7ee490 | 971 | { |
667d3740 VK |
972 | return __vmbus_recvpacket(channel, buffer, bufferlen, |
973 | buffer_actual_len, requestid, true); | |
3e7ee490 | 974 | } |
adaee6bd | 975 | EXPORT_SYMBOL_GPL(vmbus_recvpacket_raw); |