1 /* 2 * 3 * Intel Management Engine Interface (Intel MEI) Linux driver 4 * Copyright (c) 2003-2012, Intel Corporation. 5 * 6 * This program is free software; you can redistribute it and/or modify it 7 * under the terms and conditions of the GNU General Public License, 8 * version 2, as published by the Free Software Foundation. 9 * 10 * This program is distributed in the hope it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 13 * more details. 14 * 15 */ 16 17 18 #include <linux/export.h> 19 #include <linux/kthread.h> 20 #include <linux/interrupt.h> 21 #include <linux/fs.h> 22 #include <linux/jiffies.h> 23 #include <linux/slab.h> 24 #include <linux/pm_runtime.h> 25 26 #include <linux/mei.h> 27 28 #include "mei_dev.h" 29 #include "hbm.h" 30 #include "client.h" 31 32 33 /** 34 * mei_irq_compl_handler - dispatch complete handlers 35 * for the completed callbacks 36 * 37 * @dev: mei device 38 * @cmpl_list: list of completed cbs 39 */ 40 void mei_irq_compl_handler(struct mei_device *dev, struct list_head *cmpl_list) 41 { 42 struct mei_cl_cb *cb, *next; 43 struct mei_cl *cl; 44 45 list_for_each_entry_safe(cb, next, cmpl_list, list) { 46 cl = cb->cl; 47 list_del_init(&cb->list); 48 49 dev_dbg(dev->dev, "completing call back.\n"); 50 if (cl == &dev->iamthif_cl) 51 mei_amthif_complete(cl, cb); 52 else 53 mei_cl_complete(cl, cb); 54 } 55 } 56 EXPORT_SYMBOL_GPL(mei_irq_compl_handler); 57 58 /** 59 * mei_cl_hbm_equal - check if hbm is addressed to the client 60 * 61 * @cl: host client 62 * @mei_hdr: header of mei client message 63 * 64 * Return: true if matches, false otherwise 65 */ 66 static inline int mei_cl_hbm_equal(struct mei_cl *cl, 67 struct mei_msg_hdr *mei_hdr) 68 { 69 return mei_cl_host_addr(cl) == mei_hdr->host_addr && 70 mei_cl_me_id(cl) == mei_hdr->me_addr; 71 } 72 73 /** 74 * mei_irq_discard_msg - discard received message 75 * 76 * @dev: mei device 77 * @hdr: message header 78 */ 79 void mei_irq_discard_msg(struct mei_device *dev, struct mei_msg_hdr *hdr) 80 { 81 /* 82 * no need to check for size as it is guarantied 83 * that length fits into rd_msg_buf 84 */ 85 mei_read_slots(dev, dev->rd_msg_buf, hdr->length); 86 dev_dbg(dev->dev, "discarding message " MEI_HDR_FMT "\n", 87 MEI_HDR_PRM(hdr)); 88 } 89 90 /** 91 * mei_cl_irq_read_msg - process client message 92 * 93 * @cl: reading client 94 * @mei_hdr: header of mei client message 95 * @cmpl_list: completion list 96 * 97 * Return: always 0 98 */ 99 int mei_cl_irq_read_msg(struct mei_cl *cl, 100 struct mei_msg_hdr *mei_hdr, 101 struct list_head *cmpl_list) 102 { 103 struct mei_device *dev = cl->dev; 104 struct mei_cl_cb *cb; 105 size_t buf_sz; 106 107 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); 108 if (!cb) { 109 if (!mei_cl_is_fixed_address(cl)) { 110 cl_err(dev, cl, "pending read cb not found\n"); 111 goto discard; 112 } 113 cb = mei_cl_alloc_cb(cl, mei_cl_mtu(cl), MEI_FOP_READ, cl->fp); 114 if (!cb) 115 goto discard; 116 list_add_tail(&cb->list, &cl->rd_pending); 117 } 118 119 if (!mei_cl_is_connected(cl)) { 120 cl_dbg(dev, cl, "not connected\n"); 121 cb->status = -ENODEV; 122 goto discard; 123 } 124 125 buf_sz = mei_hdr->length + cb->buf_idx; 126 /* catch for integer overflow */ 127 if (buf_sz < cb->buf_idx) { 128 cl_err(dev, cl, "message is too big len %d idx %zu\n", 129 mei_hdr->length, cb->buf_idx); 130 cb->status = -EMSGSIZE; 131 goto discard; 132 } 133 134 if (cb->buf.size < buf_sz) { 135 cl_dbg(dev, cl, "message overflow. size %zu len %d idx %zu\n", 136 cb->buf.size, mei_hdr->length, cb->buf_idx); 137 cb->status = -EMSGSIZE; 138 goto discard; 139 } 140 141 mei_read_slots(dev, cb->buf.data + cb->buf_idx, mei_hdr->length); 142 143 cb->buf_idx += mei_hdr->length; 144 145 if (mei_hdr->msg_complete) { 146 cl_dbg(dev, cl, "completed read length = %zu\n", cb->buf_idx); 147 list_move_tail(&cb->list, cmpl_list); 148 } else { 149 pm_runtime_mark_last_busy(dev->dev); 150 pm_request_autosuspend(dev->dev); 151 } 152 153 return 0; 154 155 discard: 156 if (cb) 157 list_move_tail(&cb->list, cmpl_list); 158 mei_irq_discard_msg(dev, mei_hdr); 159 return 0; 160 } 161 162 /** 163 * mei_cl_irq_disconnect_rsp - send disconnection response message 164 * 165 * @cl: client 166 * @cb: callback block. 167 * @cmpl_list: complete list. 168 * 169 * Return: 0, OK; otherwise, error. 170 */ 171 static int mei_cl_irq_disconnect_rsp(struct mei_cl *cl, struct mei_cl_cb *cb, 172 struct list_head *cmpl_list) 173 { 174 struct mei_device *dev = cl->dev; 175 u32 msg_slots; 176 int slots; 177 int ret; 178 179 slots = mei_hbuf_empty_slots(dev); 180 msg_slots = mei_data2slots(sizeof(struct hbm_client_connect_response)); 181 182 if (slots < msg_slots) 183 return -EMSGSIZE; 184 185 ret = mei_hbm_cl_disconnect_rsp(dev, cl); 186 list_move_tail(&cb->list, cmpl_list); 187 188 return ret; 189 } 190 191 /** 192 * mei_cl_irq_read - processes client read related operation from the 193 * interrupt thread context - request for flow control credits 194 * 195 * @cl: client 196 * @cb: callback block. 197 * @cmpl_list: complete list. 198 * 199 * Return: 0, OK; otherwise, error. 200 */ 201 static int mei_cl_irq_read(struct mei_cl *cl, struct mei_cl_cb *cb, 202 struct list_head *cmpl_list) 203 { 204 struct mei_device *dev = cl->dev; 205 u32 msg_slots; 206 int slots; 207 int ret; 208 209 if (!list_empty(&cl->rd_pending)) 210 return 0; 211 212 msg_slots = mei_data2slots(sizeof(struct hbm_flow_control)); 213 slots = mei_hbuf_empty_slots(dev); 214 215 if (slots < msg_slots) 216 return -EMSGSIZE; 217 218 ret = mei_hbm_cl_flow_control_req(dev, cl); 219 if (ret) { 220 cl->status = ret; 221 cb->buf_idx = 0; 222 list_move_tail(&cb->list, cmpl_list); 223 return ret; 224 } 225 226 list_move_tail(&cb->list, &cl->rd_pending); 227 228 return 0; 229 } 230 231 static inline bool hdr_is_hbm(struct mei_msg_hdr *mei_hdr) 232 { 233 return mei_hdr->host_addr == 0 && mei_hdr->me_addr == 0; 234 } 235 236 static inline bool hdr_is_fixed(struct mei_msg_hdr *mei_hdr) 237 { 238 return mei_hdr->host_addr == 0 && mei_hdr->me_addr != 0; 239 } 240 241 /** 242 * mei_irq_read_handler - bottom half read routine after ISR to 243 * handle the read processing. 244 * 245 * @dev: the device structure 246 * @cmpl_list: An instance of our list structure 247 * @slots: slots to read. 248 * 249 * Return: 0 on success, <0 on failure. 250 */ 251 int mei_irq_read_handler(struct mei_device *dev, 252 struct list_head *cmpl_list, s32 *slots) 253 { 254 struct mei_msg_hdr *mei_hdr; 255 struct mei_cl *cl; 256 int ret; 257 258 if (!dev->rd_msg_hdr) { 259 dev->rd_msg_hdr = mei_read_hdr(dev); 260 (*slots)--; 261 dev_dbg(dev->dev, "slots =%08x.\n", *slots); 262 } 263 mei_hdr = (struct mei_msg_hdr *) &dev->rd_msg_hdr; 264 dev_dbg(dev->dev, MEI_HDR_FMT, MEI_HDR_PRM(mei_hdr)); 265 266 if (mei_hdr->reserved || !dev->rd_msg_hdr) { 267 dev_err(dev->dev, "corrupted message header 0x%08X\n", 268 dev->rd_msg_hdr); 269 ret = -EBADMSG; 270 goto end; 271 } 272 273 if (mei_slots2data(*slots) < mei_hdr->length) { 274 dev_err(dev->dev, "less data available than length=%08x.\n", 275 *slots); 276 /* we can't read the message */ 277 ret = -ENODATA; 278 goto end; 279 } 280 281 /* HBM message */ 282 if (hdr_is_hbm(mei_hdr)) { 283 ret = mei_hbm_dispatch(dev, mei_hdr); 284 if (ret) { 285 dev_dbg(dev->dev, "mei_hbm_dispatch failed ret = %d\n", 286 ret); 287 goto end; 288 } 289 goto reset_slots; 290 } 291 292 /* find recipient cl */ 293 list_for_each_entry(cl, &dev->file_list, link) { 294 if (mei_cl_hbm_equal(cl, mei_hdr)) { 295 cl_dbg(dev, cl, "got a message\n"); 296 break; 297 } 298 } 299 300 /* if no recipient cl was found we assume corrupted header */ 301 if (&cl->link == &dev->file_list) { 302 /* A message for not connected fixed address clients 303 * should be silently discarded 304 */ 305 if (hdr_is_fixed(mei_hdr)) { 306 mei_irq_discard_msg(dev, mei_hdr); 307 ret = 0; 308 goto reset_slots; 309 } 310 dev_err(dev->dev, "no destination client found 0x%08X\n", 311 dev->rd_msg_hdr); 312 ret = -EBADMSG; 313 goto end; 314 } 315 316 if (cl == &dev->iamthif_cl) { 317 ret = mei_amthif_irq_read_msg(cl, mei_hdr, cmpl_list); 318 } else { 319 ret = mei_cl_irq_read_msg(cl, mei_hdr, cmpl_list); 320 } 321 322 323 reset_slots: 324 /* reset the number of slots and header */ 325 *slots = mei_count_full_read_slots(dev); 326 dev->rd_msg_hdr = 0; 327 328 if (*slots == -EOVERFLOW) { 329 /* overflow - reset */ 330 dev_err(dev->dev, "resetting due to slots overflow.\n"); 331 /* set the event since message has been read */ 332 ret = -ERANGE; 333 goto end; 334 } 335 end: 336 return ret; 337 } 338 EXPORT_SYMBOL_GPL(mei_irq_read_handler); 339 340 341 /** 342 * mei_irq_write_handler - dispatch write requests 343 * after irq received 344 * 345 * @dev: the device structure 346 * @cmpl_list: An instance of our list structure 347 * 348 * Return: 0 on success, <0 on failure. 349 */ 350 int mei_irq_write_handler(struct mei_device *dev, struct list_head *cmpl_list) 351 { 352 353 struct mei_cl *cl; 354 struct mei_cl_cb *cb, *next; 355 s32 slots; 356 int ret; 357 358 359 if (!mei_hbuf_acquire(dev)) 360 return 0; 361 362 slots = mei_hbuf_empty_slots(dev); 363 if (slots <= 0) 364 return -EMSGSIZE; 365 366 /* complete all waiting for write CB */ 367 dev_dbg(dev->dev, "complete all waiting for write cb.\n"); 368 369 list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) { 370 cl = cb->cl; 371 372 cl->status = 0; 373 cl_dbg(dev, cl, "MEI WRITE COMPLETE\n"); 374 cl->writing_state = MEI_WRITE_COMPLETE; 375 list_move_tail(&cb->list, cmpl_list); 376 } 377 378 /* complete control write list CB */ 379 dev_dbg(dev->dev, "complete control write list cb.\n"); 380 list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) { 381 cl = cb->cl; 382 switch (cb->fop_type) { 383 case MEI_FOP_DISCONNECT: 384 /* send disconnect message */ 385 ret = mei_cl_irq_disconnect(cl, cb, cmpl_list); 386 if (ret) 387 return ret; 388 389 break; 390 case MEI_FOP_READ: 391 /* send flow control message */ 392 ret = mei_cl_irq_read(cl, cb, cmpl_list); 393 if (ret) 394 return ret; 395 396 break; 397 case MEI_FOP_CONNECT: 398 /* connect message */ 399 ret = mei_cl_irq_connect(cl, cb, cmpl_list); 400 if (ret) 401 return ret; 402 403 break; 404 case MEI_FOP_DISCONNECT_RSP: 405 /* send disconnect resp */ 406 ret = mei_cl_irq_disconnect_rsp(cl, cb, cmpl_list); 407 if (ret) 408 return ret; 409 break; 410 411 case MEI_FOP_NOTIFY_START: 412 case MEI_FOP_NOTIFY_STOP: 413 ret = mei_cl_irq_notify(cl, cb, cmpl_list); 414 if (ret) 415 return ret; 416 break; 417 default: 418 BUG(); 419 } 420 421 } 422 /* complete write list CB */ 423 dev_dbg(dev->dev, "complete write list cb.\n"); 424 list_for_each_entry_safe(cb, next, &dev->write_list, list) { 425 cl = cb->cl; 426 if (cl == &dev->iamthif_cl) 427 ret = mei_amthif_irq_write(cl, cb, cmpl_list); 428 else 429 ret = mei_cl_irq_write(cl, cb, cmpl_list); 430 if (ret) 431 return ret; 432 } 433 return 0; 434 } 435 EXPORT_SYMBOL_GPL(mei_irq_write_handler); 436 437 438 /** 439 * mei_connect_timeout - connect/disconnect timeouts 440 * 441 * @cl: host client 442 */ 443 static void mei_connect_timeout(struct mei_cl *cl) 444 { 445 struct mei_device *dev = cl->dev; 446 447 if (cl->state == MEI_FILE_CONNECTING) { 448 if (dev->hbm_f_dot_supported) { 449 cl->state = MEI_FILE_DISCONNECT_REQUIRED; 450 wake_up(&cl->wait); 451 return; 452 } 453 } 454 mei_reset(dev); 455 } 456 457 #define MEI_STALL_TIMER_FREQ (2 * HZ) 458 /** 459 * mei_schedule_stall_timer - re-arm stall_timer work 460 * 461 * Schedule stall timer 462 * 463 * @dev: the device structure 464 */ 465 void mei_schedule_stall_timer(struct mei_device *dev) 466 { 467 schedule_delayed_work(&dev->timer_work, MEI_STALL_TIMER_FREQ); 468 } 469 470 /** 471 * mei_timer - timer function. 472 * 473 * @work: pointer to the work_struct structure 474 * 475 */ 476 void mei_timer(struct work_struct *work) 477 { 478 struct mei_cl *cl; 479 struct mei_device *dev = container_of(work, 480 struct mei_device, timer_work.work); 481 bool reschedule_timer = false; 482 483 mutex_lock(&dev->device_lock); 484 485 /* Catch interrupt stalls during HBM init handshake */ 486 if (dev->dev_state == MEI_DEV_INIT_CLIENTS && 487 dev->hbm_state != MEI_HBM_IDLE) { 488 489 if (dev->init_clients_timer) { 490 if (--dev->init_clients_timer == 0) { 491 dev_err(dev->dev, "timer: init clients timeout hbm_state = %d.\n", 492 dev->hbm_state); 493 mei_reset(dev); 494 goto out; 495 } 496 reschedule_timer = true; 497 } 498 } 499 500 if (dev->dev_state != MEI_DEV_ENABLED) 501 goto out; 502 503 /*** connect/disconnect timeouts ***/ 504 list_for_each_entry(cl, &dev->file_list, link) { 505 if (cl->timer_count) { 506 if (--cl->timer_count == 0) { 507 dev_err(dev->dev, "timer: connect/disconnect timeout.\n"); 508 mei_connect_timeout(cl); 509 goto out; 510 } 511 reschedule_timer = true; 512 } 513 } 514 515 if (!mei_cl_is_connected(&dev->iamthif_cl)) 516 goto out; 517 518 if (dev->iamthif_stall_timer) { 519 if (--dev->iamthif_stall_timer == 0) { 520 dev_err(dev->dev, "timer: amthif hanged.\n"); 521 mei_reset(dev); 522 523 mei_amthif_run_next_cmd(dev); 524 goto out; 525 } 526 reschedule_timer = true; 527 } 528 529 out: 530 if (dev->dev_state != MEI_DEV_DISABLED && reschedule_timer) 531 mei_schedule_stall_timer(dev); 532 533 mutex_unlock(&dev->device_lock); 534 } 535