Print this page
XXXX introduce drv_sectohz
Split |
Close |
Expand all |
Collapse all |
--- old/usr/src/uts/common/io/comstar/port/iscsit/iscsit.c
+++ new/usr/src/uts/common/io/comstar/port/iscsit/iscsit.c
1 1 /*
2 2 * CDDL HEADER START
3 3 *
4 4 * The contents of this file are subject to the terms of the
5 5 * Common Development and Distribution License (the "License").
6 6 * You may not use this file except in compliance with the License.
7 7 *
8 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 9 * or http://www.opensolaris.org/os/licensing.
10 10 * See the License for the specific language governing permissions
11 11 * and limitations under the License.
12 12 *
13 13 * When distributing Covered Code, include this CDDL HEADER in each
14 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 15 * If applicable, add the following below this CDDL HEADER, with the
16 16 * fields enclosed by brackets "[]" replaced with your own identifying
17 17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 18 *
19 19 * CDDL HEADER END
20 20 */
21 21 /*
22 22 * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
23 23 *
24 24 * Copyright 2014 Nexenta Systems, Inc. All rights reserved.
25 25 */
26 26
27 27 #include <sys/cpuvar.h>
28 28 #include <sys/types.h>
29 29 #include <sys/conf.h>
30 30 #include <sys/stat.h>
31 31 #include <sys/file.h>
32 32 #include <sys/ddi.h>
33 33 #include <sys/sunddi.h>
34 34 #include <sys/modctl.h>
35 35 #include <sys/sysmacros.h>
36 36 #include <sys/socket.h>
37 37 #include <sys/strsubr.h>
38 38 #include <sys/nvpair.h>
39 39
40 40 #include <sys/stmf.h>
41 41 #include <sys/stmf_ioctl.h>
42 42 #include <sys/portif.h>
43 43 #include <sys/idm/idm.h>
44 44 #include <sys/idm/idm_conn_sm.h>
45 45
46 46 #include "iscsit_isns.h"
47 47 #include "iscsit.h"
48 48
49 49 #define ISCSIT_VERSION BUILD_DATE "-1.18dev"
50 50 #define ISCSIT_NAME_VERSION "COMSTAR ISCSIT v" ISCSIT_VERSION
51 51
52 52 /*
53 53 * DDI entry points.
54 54 */
55 55 static int iscsit_drv_attach(dev_info_t *, ddi_attach_cmd_t);
56 56 static int iscsit_drv_detach(dev_info_t *, ddi_detach_cmd_t);
57 57 static int iscsit_drv_getinfo(dev_info_t *, ddi_info_cmd_t, void *, void **);
58 58 static int iscsit_drv_open(dev_t *, int, int, cred_t *);
59 59 static int iscsit_drv_close(dev_t, int, int, cred_t *);
60 60 static boolean_t iscsit_drv_busy(void);
61 61 static int iscsit_drv_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
62 62
63 63 extern struct mod_ops mod_miscops;
64 64
65 65
66 66 static struct cb_ops iscsit_cb_ops = {
67 67 iscsit_drv_open, /* cb_open */
68 68 iscsit_drv_close, /* cb_close */
69 69 nodev, /* cb_strategy */
70 70 nodev, /* cb_print */
71 71 nodev, /* cb_dump */
72 72 nodev, /* cb_read */
73 73 nodev, /* cb_write */
74 74 iscsit_drv_ioctl, /* cb_ioctl */
75 75 nodev, /* cb_devmap */
76 76 nodev, /* cb_mmap */
77 77 nodev, /* cb_segmap */
78 78 nochpoll, /* cb_chpoll */
79 79 ddi_prop_op, /* cb_prop_op */
80 80 NULL, /* cb_streamtab */
81 81 D_MP, /* cb_flag */
82 82 CB_REV, /* cb_rev */
83 83 nodev, /* cb_aread */
84 84 nodev, /* cb_awrite */
85 85 };
86 86
87 87 static struct dev_ops iscsit_dev_ops = {
88 88 DEVO_REV, /* devo_rev */
89 89 0, /* devo_refcnt */
90 90 iscsit_drv_getinfo, /* devo_getinfo */
91 91 nulldev, /* devo_identify */
92 92 nulldev, /* devo_probe */
93 93 iscsit_drv_attach, /* devo_attach */
94 94 iscsit_drv_detach, /* devo_detach */
95 95 nodev, /* devo_reset */
96 96 &iscsit_cb_ops, /* devo_cb_ops */
97 97 NULL, /* devo_bus_ops */
98 98 NULL, /* devo_power */
99 99 ddi_quiesce_not_needed, /* quiesce */
100 100 };
101 101
102 102 static struct modldrv modldrv = {
103 103 &mod_driverops,
104 104 "iSCSI Target",
105 105 &iscsit_dev_ops,
106 106 };
107 107
108 108 static struct modlinkage modlinkage = {
109 109 MODREV_1,
110 110 &modldrv,
111 111 NULL,
112 112 };
113 113
114 114
115 115 iscsit_global_t iscsit_global;
116 116
117 117 kmem_cache_t *iscsit_status_pdu_cache;
118 118
119 119 boolean_t iscsit_sm_logging = B_FALSE;
120 120
121 121 kmutex_t login_sm_session_mutex;
122 122
123 123 static idm_status_t iscsit_init(dev_info_t *dip);
124 124 static idm_status_t iscsit_enable_svc(iscsit_hostinfo_t *hostinfo);
125 125 static void iscsit_disable_svc(void);
126 126
127 127 static int
128 128 iscsit_check_cmdsn_and_queue(idm_pdu_t *rx_pdu);
129 129
130 130 static void
131 131 iscsit_add_pdu_to_queue(iscsit_sess_t *ist, idm_pdu_t *rx_pdu);
132 132
133 133 static idm_pdu_t *
134 134 iscsit_remove_pdu_from_queue(iscsit_sess_t *ist, uint32_t cmdsn);
135 135
136 136 static void
137 137 iscsit_process_pdu_in_queue(iscsit_sess_t *ist);
138 138
139 139 static void
140 140 iscsit_rxpdu_queue_monitor_session(iscsit_sess_t *ist);
141 141
142 142 static void
143 143 iscsit_rxpdu_queue_monitor(void *arg);
144 144
145 145 static void
146 146 iscsit_post_staged_pdu(idm_pdu_t *rx_pdu);
147 147
148 148 static void
149 149 iscsit_post_scsi_cmd(idm_conn_t *ic, idm_pdu_t *rx_pdu);
150 150
151 151 static void
152 152 iscsit_op_scsi_task_mgmt(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
153 153
154 154 static void
155 155 iscsit_pdu_op_noop(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
156 156
157 157 static void
158 158 iscsit_pdu_op_login_cmd(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
159 159
160 160 void
161 161 iscsit_pdu_op_text_cmd(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
162 162
163 163 static void
164 164 iscsit_pdu_op_logout_cmd(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
165 165
166 166 int iscsit_cmd_window();
167 167
168 168 static int
169 169 iscsit_sna_lt(uint32_t sn1, uint32_t sn2);
170 170
171 171 void
172 172 iscsit_set_cmdsn(iscsit_conn_t *ict, idm_pdu_t *rx_pdu);
173 173
174 174 static void
175 175 iscsit_deferred_dispatch(idm_pdu_t *rx_pdu);
176 176
177 177 static void
178 178 iscsit_deferred(void *rx_pdu_void);
179 179
180 180 static idm_status_t
181 181 iscsit_conn_accept(idm_conn_t *ic);
182 182
183 183 static idm_status_t
184 184 iscsit_ffp_enabled(idm_conn_t *ic);
185 185
186 186 static idm_status_t
187 187 iscsit_ffp_disabled(idm_conn_t *ic, idm_ffp_disable_t disable_class);
188 188
189 189 static idm_status_t
190 190 iscsit_conn_lost(idm_conn_t *ic);
191 191
192 192 static idm_status_t
193 193 iscsit_conn_destroy(idm_conn_t *ic);
194 194
195 195 static stmf_data_buf_t *
196 196 iscsit_dbuf_alloc(scsi_task_t *task, uint32_t size, uint32_t *pminsize,
197 197 uint32_t flags);
198 198
199 199 static void
200 200 iscsit_dbuf_free(stmf_dbuf_store_t *ds, stmf_data_buf_t *dbuf);
201 201
202 202 static void
203 203 iscsit_buf_xfer_cb(idm_buf_t *idb, idm_status_t status);
204 204
205 205 static void
206 206 iscsit_send_good_status_done(idm_pdu_t *pdu, idm_status_t status);
207 207
208 208 static void
209 209 iscsit_send_status_done(idm_pdu_t *pdu, idm_status_t status);
210 210
211 211 static stmf_status_t
212 212 iscsit_idm_to_stmf(idm_status_t idmrc);
213 213
214 214 static iscsit_task_t *
215 215 iscsit_task_alloc(iscsit_conn_t *ict);
216 216
217 217 static void
218 218 iscsit_task_free(iscsit_task_t *itask);
219 219
220 220 static iscsit_task_t *
221 221 iscsit_tm_task_alloc(iscsit_conn_t *ict);
222 222
223 223 static void
224 224 iscsit_tm_task_free(iscsit_task_t *itask);
225 225
226 226 static idm_status_t
227 227 iscsit_task_start(iscsit_task_t *itask);
228 228
229 229 static void
230 230 iscsit_task_done(iscsit_task_t *itask);
231 231
232 232 static int
233 233 iscsit_status_pdu_constructor(void *pdu_void, void *arg, int flags);
234 234
235 235 static void
236 236 iscsit_pp_cb(struct stmf_port_provider *pp, int cmd, void *arg, uint32_t flags);
237 237
238 238 static it_cfg_status_t
239 239 iscsit_config_merge(it_config_t *cfg);
240 240
241 241 static idm_status_t
242 242 iscsit_login_fail(idm_conn_t *ic);
243 243
244 244 static boolean_t iscsit_cmdsn_in_window(iscsit_conn_t *ict, uint32_t cmdsn);
245 245 static void iscsit_send_direct_scsi_resp(iscsit_conn_t *ict, idm_pdu_t *rx_pdu,
246 246 uint8_t response, uint8_t cmd_status);
247 247 static void iscsit_send_task_mgmt_resp(idm_pdu_t *tm_resp_pdu,
248 248 uint8_t tm_status);
249 249
250 250 /*
251 251 * MC/S: Out-of-order commands are staged on a session-wide wait
252 252 * queue until a system-tunable threshold is reached. A separate
253 253 * thread is used to scan the staging queue on all the session,
254 254 * If a delayed PDU does not arrive within a timeout, the target
255 255 * will advance to the staged PDU that is next in sequence, skipping
256 256 * over the missing PDU(s) to go past a hole in the sequence.
257 257 */
258 258 volatile int rxpdu_queue_threshold = ISCSIT_RXPDU_QUEUE_THRESHOLD;
259 259
260 260 static kmutex_t iscsit_rxpdu_queue_monitor_mutex;
261 261 kthread_t *iscsit_rxpdu_queue_monitor_thr_id;
262 262 static kt_did_t iscsit_rxpdu_queue_monitor_thr_did;
263 263 static boolean_t iscsit_rxpdu_queue_monitor_thr_running;
264 264 static kcondvar_t iscsit_rxpdu_queue_monitor_cv;
265 265
266 266 int
267 267 _init(void)
268 268 {
269 269 int rc;
270 270
271 271 rw_init(&iscsit_global.global_rwlock, NULL, RW_DRIVER, NULL);
272 272 mutex_init(&iscsit_global.global_state_mutex, NULL,
273 273 MUTEX_DRIVER, NULL);
274 274 iscsit_global.global_svc_state = ISE_DETACHED;
275 275
276 276 mutex_init(&iscsit_rxpdu_queue_monitor_mutex, NULL,
277 277 MUTEX_DRIVER, NULL);
278 278 mutex_init(&login_sm_session_mutex, NULL, MUTEX_DRIVER, NULL);
279 279 iscsit_rxpdu_queue_monitor_thr_id = NULL;
280 280 iscsit_rxpdu_queue_monitor_thr_running = B_FALSE;
281 281 cv_init(&iscsit_rxpdu_queue_monitor_cv, NULL, CV_DEFAULT, NULL);
282 282
283 283 if ((rc = mod_install(&modlinkage)) != 0) {
284 284 mutex_destroy(&iscsit_global.global_state_mutex);
285 285 rw_destroy(&iscsit_global.global_rwlock);
286 286 return (rc);
287 287 }
288 288
289 289 return (rc);
290 290 }
291 291
292 292 int
293 293 _info(struct modinfo *modinfop)
294 294 {
295 295 return (mod_info(&modlinkage, modinfop));
296 296 }
297 297
298 298 int
299 299 _fini(void)
300 300 {
301 301 int rc;
302 302
303 303 rc = mod_remove(&modlinkage);
304 304
305 305 if (rc == 0) {
306 306 mutex_destroy(&iscsit_rxpdu_queue_monitor_mutex);
307 307 mutex_destroy(&login_sm_session_mutex);
308 308 cv_destroy(&iscsit_rxpdu_queue_monitor_cv);
309 309 mutex_destroy(&iscsit_global.global_state_mutex);
310 310 rw_destroy(&iscsit_global.global_rwlock);
311 311 }
312 312
313 313 return (rc);
314 314 }
315 315
316 316 /*
317 317 * DDI entry points.
318 318 */
319 319
320 320 /* ARGSUSED */
321 321 static int
322 322 iscsit_drv_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg,
323 323 void **result)
324 324 {
325 325 ulong_t instance = getminor((dev_t)arg);
326 326
327 327 switch (cmd) {
328 328 case DDI_INFO_DEVT2DEVINFO:
329 329 *result = iscsit_global.global_dip;
330 330 return (DDI_SUCCESS);
331 331
332 332 case DDI_INFO_DEVT2INSTANCE:
333 333 *result = (void *)instance;
334 334 return (DDI_SUCCESS);
335 335
336 336 default:
337 337 break;
338 338 }
339 339
340 340 return (DDI_FAILURE);
341 341 }
342 342
343 343 static int
344 344 iscsit_drv_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
345 345 {
346 346 if (cmd != DDI_ATTACH) {
347 347 return (DDI_FAILURE);
348 348 }
349 349
350 350 if (ddi_get_instance(dip) != 0) {
351 351 /* we only allow instance 0 to attach */
352 352 return (DDI_FAILURE);
353 353 }
354 354
355 355 /* create the minor node */
356 356 if (ddi_create_minor_node(dip, ISCSIT_MODNAME, S_IFCHR, 0,
357 357 DDI_PSEUDO, 0) != DDI_SUCCESS) {
358 358 cmn_err(CE_WARN, "iscsit_drv_attach: "
359 359 "failed creating minor node");
360 360 return (DDI_FAILURE);
361 361 }
362 362
363 363 if (iscsit_init(dip) != IDM_STATUS_SUCCESS) {
364 364 cmn_err(CE_WARN, "iscsit_drv_attach: "
365 365 "failed to initialize");
366 366 ddi_remove_minor_node(dip, NULL);
367 367 return (DDI_FAILURE);
368 368 }
369 369
370 370 iscsit_global.global_svc_state = ISE_DISABLED;
371 371 iscsit_global.global_dip = dip;
372 372
373 373 return (DDI_SUCCESS);
374 374 }
375 375
376 376 /*ARGSUSED*/
377 377 static int
378 378 iscsit_drv_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
379 379 {
380 380 if (cmd != DDI_DETACH)
381 381 return (DDI_FAILURE);
382 382
383 383 /*
384 384 * drv_detach is called in a context that owns the
385 385 * device node for the /dev/pseudo device. If this thread blocks
386 386 * for any resource, other threads that need the /dev/pseudo device
387 387 * may end up in a deadlock with this thread.Hence, we use a
388 388 * separate lock just for the structures that drv_detach needs
389 389 * to access.
390 390 */
391 391 mutex_enter(&iscsit_global.global_state_mutex);
392 392 if (iscsit_drv_busy()) {
393 393 mutex_exit(&iscsit_global.global_state_mutex);
394 394 return (EBUSY);
395 395 }
396 396
397 397 iscsit_global.global_dip = NULL;
398 398 ddi_remove_minor_node(dip, NULL);
399 399
400 400 ldi_ident_release(iscsit_global.global_li);
401 401 iscsit_global.global_svc_state = ISE_DETACHED;
402 402
403 403 mutex_exit(&iscsit_global.global_state_mutex);
404 404
405 405 return (DDI_SUCCESS);
406 406 }
407 407
408 408 /*ARGSUSED*/
409 409 static int
410 410 iscsit_drv_open(dev_t *devp, int flag, int otyp, cred_t *credp)
411 411 {
412 412 return (0);
413 413 }
414 414
415 415 /* ARGSUSED */
416 416 static int
417 417 iscsit_drv_close(dev_t dev, int flag, int otyp, cred_t *credp)
418 418 {
419 419 return (0);
420 420 }
421 421
422 422 static boolean_t
423 423 iscsit_drv_busy(void)
424 424 {
425 425 ASSERT(MUTEX_HELD(&iscsit_global.global_state_mutex));
426 426
427 427 switch (iscsit_global.global_svc_state) {
428 428 case ISE_DISABLED:
429 429 case ISE_DETACHED:
430 430 return (B_FALSE);
431 431 default:
432 432 return (B_TRUE);
433 433 }
434 434 /* NOTREACHED */
435 435 }
436 436
437 437 /* ARGSUSED */
438 438 static int
439 439 iscsit_drv_ioctl(dev_t drv, int cmd, intptr_t argp, int flag, cred_t *cred,
440 440 int *retval)
441 441 {
442 442 iscsit_ioc_set_config_t setcfg;
443 443 iscsit_ioc_set_config32_t setcfg32;
444 444 char *cfg_pnvlist = NULL;
445 445 nvlist_t *cfg_nvlist = NULL;
446 446 it_config_t *cfg = NULL;
447 447 idm_status_t idmrc;
448 448 int rc = 0;
449 449
450 450 if (drv_priv(cred) != 0) {
451 451 return (EPERM);
452 452 }
453 453
454 454 mutex_enter(&iscsit_global.global_state_mutex);
455 455
456 456 /*
457 457 * Validate ioctl requests against global service state
458 458 */
459 459 switch (iscsit_global.global_svc_state) {
460 460 case ISE_ENABLED:
461 461 if (cmd == ISCSIT_IOC_DISABLE_SVC) {
462 462 iscsit_global.global_svc_state = ISE_DISABLING;
463 463 } else if (cmd == ISCSIT_IOC_ENABLE_SVC) {
464 464 /* Already enabled */
465 465 mutex_exit(&iscsit_global.global_state_mutex);
466 466 return (0);
467 467 } else {
468 468 iscsit_global.global_svc_state = ISE_BUSY;
469 469 }
470 470 break;
471 471 case ISE_DISABLED:
472 472 if (cmd == ISCSIT_IOC_ENABLE_SVC) {
473 473 iscsit_global.global_svc_state = ISE_ENABLING;
474 474 } else if (cmd == ISCSIT_IOC_DISABLE_SVC) {
475 475 /* Already disabled */
476 476 mutex_exit(&iscsit_global.global_state_mutex);
477 477 return (0);
478 478 } else {
479 479 rc = EFAULT;
480 480 }
481 481 break;
482 482 case ISE_BUSY:
483 483 case ISE_ENABLING:
484 484 case ISE_DISABLING:
485 485 rc = EAGAIN;
486 486 break;
487 487 case ISE_DETACHED:
488 488 default:
489 489 rc = EFAULT;
490 490 break;
491 491 }
492 492
493 493 mutex_exit(&iscsit_global.global_state_mutex);
494 494 if (rc != 0)
495 495 return (rc);
496 496
497 497 /* Handle ioctl request (enable/disable have already been handled) */
498 498 switch (cmd) {
499 499 case ISCSIT_IOC_SET_CONFIG:
500 500 /* Any errors must set state back to ISE_ENABLED */
501 501 switch (ddi_model_convert_from(flag & FMODELS)) {
502 502 case DDI_MODEL_ILP32:
503 503 if (ddi_copyin((void *)argp, &setcfg32,
504 504 sizeof (iscsit_ioc_set_config32_t), flag) != 0) {
505 505 rc = EFAULT;
506 506 goto cleanup;
507 507 }
508 508
509 509 setcfg.set_cfg_pnvlist =
510 510 (char *)((uintptr_t)setcfg32.set_cfg_pnvlist);
511 511 setcfg.set_cfg_vers = setcfg32.set_cfg_vers;
512 512 setcfg.set_cfg_pnvlist_len =
513 513 setcfg32.set_cfg_pnvlist_len;
514 514 break;
515 515 case DDI_MODEL_NONE:
516 516 if (ddi_copyin((void *)argp, &setcfg,
517 517 sizeof (iscsit_ioc_set_config_t), flag) != 0) {
518 518 rc = EFAULT;
519 519 goto cleanup;
520 520 }
521 521 break;
522 522 default:
523 523 rc = EFAULT;
524 524 goto cleanup;
525 525 }
526 526
527 527 /* Check API version */
528 528 if (setcfg.set_cfg_vers != ISCSIT_API_VERS0) {
529 529 rc = EINVAL;
530 530 goto cleanup;
531 531 }
532 532
533 533 /* Config is in packed nvlist format so unpack it */
534 534 cfg_pnvlist = kmem_alloc(setcfg.set_cfg_pnvlist_len,
535 535 KM_SLEEP);
536 536 ASSERT(cfg_pnvlist != NULL);
537 537
538 538 if (ddi_copyin(setcfg.set_cfg_pnvlist, cfg_pnvlist,
539 539 setcfg.set_cfg_pnvlist_len, flag) != 0) {
540 540 rc = EFAULT;
541 541 goto cleanup;
542 542 }
543 543
544 544 rc = nvlist_unpack(cfg_pnvlist, setcfg.set_cfg_pnvlist_len,
545 545 &cfg_nvlist, KM_SLEEP);
546 546 if (rc != 0) {
547 547 goto cleanup;
548 548 }
549 549
550 550 /* Translate nvlist */
551 551 rc = it_nv_to_config(cfg_nvlist, &cfg);
552 552 if (rc != 0) {
553 553 cmn_err(CE_WARN, "Configuration is invalid");
554 554 goto cleanup;
555 555 }
556 556
557 557 /* Update config */
558 558 rc = iscsit_config_merge(cfg);
559 559 /* FALLTHROUGH */
560 560
561 561 cleanup:
562 562 if (cfg)
563 563 it_config_free_cmn(cfg);
564 564 if (cfg_pnvlist)
565 565 kmem_free(cfg_pnvlist, setcfg.set_cfg_pnvlist_len);
566 566 if (cfg_nvlist)
567 567 nvlist_free(cfg_nvlist);
568 568
569 569 /*
570 570 * Now that the reconfig is complete set our state back to
571 571 * enabled.
572 572 */
573 573 mutex_enter(&iscsit_global.global_state_mutex);
574 574 iscsit_global.global_svc_state = ISE_ENABLED;
575 575 mutex_exit(&iscsit_global.global_state_mutex);
576 576 break;
577 577 case ISCSIT_IOC_ENABLE_SVC: {
578 578 iscsit_hostinfo_t hostinfo;
579 579
580 580 if (ddi_copyin((void *)argp, &hostinfo.length,
581 581 sizeof (hostinfo.length), flag) != 0) {
582 582 mutex_enter(&iscsit_global.global_state_mutex);
583 583 iscsit_global.global_svc_state = ISE_DISABLED;
584 584 mutex_exit(&iscsit_global.global_state_mutex);
585 585 return (EFAULT);
586 586 }
587 587
588 588 if (hostinfo.length > sizeof (hostinfo.fqhn))
589 589 hostinfo.length = sizeof (hostinfo.fqhn);
590 590
591 591 if (ddi_copyin((void *)((caddr_t)argp +
592 592 sizeof (hostinfo.length)), &hostinfo.fqhn,
593 593 hostinfo.length, flag) != 0) {
594 594 mutex_enter(&iscsit_global.global_state_mutex);
595 595 iscsit_global.global_svc_state = ISE_DISABLED;
596 596 mutex_exit(&iscsit_global.global_state_mutex);
597 597 return (EFAULT);
598 598 }
599 599
600 600 idmrc = iscsit_enable_svc(&hostinfo);
601 601 mutex_enter(&iscsit_global.global_state_mutex);
602 602 if (idmrc == IDM_STATUS_SUCCESS) {
603 603 iscsit_global.global_svc_state = ISE_ENABLED;
604 604 } else {
605 605 rc = EIO;
606 606 iscsit_global.global_svc_state = ISE_DISABLED;
607 607 }
608 608 mutex_exit(&iscsit_global.global_state_mutex);
609 609 break;
610 610 }
611 611 case ISCSIT_IOC_DISABLE_SVC:
612 612 iscsit_disable_svc();
613 613 mutex_enter(&iscsit_global.global_state_mutex);
614 614 iscsit_global.global_svc_state = ISE_DISABLED;
615 615 mutex_exit(&iscsit_global.global_state_mutex);
616 616 break;
617 617
618 618 default:
619 619 rc = EINVAL;
620 620 mutex_enter(&iscsit_global.global_state_mutex);
621 621 iscsit_global.global_svc_state = ISE_ENABLED;
622 622 mutex_exit(&iscsit_global.global_state_mutex);
623 623 }
624 624
625 625 return (rc);
626 626 }
627 627
628 628 static idm_status_t
629 629 iscsit_init(dev_info_t *dip)
630 630 {
631 631 int rc;
632 632
633 633 rc = ldi_ident_from_dip(dip, &iscsit_global.global_li);
634 634 ASSERT(rc == 0); /* Failure indicates invalid argument */
635 635
636 636 iscsit_global.global_svc_state = ISE_DISABLED;
637 637
638 638 return (IDM_STATUS_SUCCESS);
639 639 }
640 640
641 641 /*
642 642 * iscsit_enable_svc
643 643 *
644 644 * registers all the configured targets and target portals with STMF
645 645 */
646 646 static idm_status_t
647 647 iscsit_enable_svc(iscsit_hostinfo_t *hostinfo)
648 648 {
649 649 stmf_port_provider_t *pp;
650 650 stmf_dbuf_store_t *dbuf_store;
651 651 boolean_t did_iscsit_isns_init;
652 652 idm_status_t retval = IDM_STATUS_SUCCESS;
653 653
654 654 ASSERT(iscsit_global.global_svc_state == ISE_ENABLING);
655 655
656 656 /*
657 657 * Make sure that can tell if we have partially allocated
658 658 * in case we need to exit and tear down anything allocated.
659 659 */
660 660 iscsit_global.global_tsih_pool = NULL;
661 661 iscsit_global.global_dbuf_store = NULL;
662 662 iscsit_status_pdu_cache = NULL;
663 663 pp = NULL;
664 664 iscsit_global.global_pp = NULL;
665 665 iscsit_global.global_default_tpg = NULL;
666 666 did_iscsit_isns_init = B_FALSE;
667 667 iscsit_global.global_dispatch_taskq = NULL;
668 668
669 669 /* Setup remaining fields in iscsit_global_t */
670 670 idm_refcnt_init(&iscsit_global.global_refcnt,
671 671 &iscsit_global);
672 672
673 673 avl_create(&iscsit_global.global_discovery_sessions,
674 674 iscsit_sess_avl_compare, sizeof (iscsit_sess_t),
675 675 offsetof(iscsit_sess_t, ist_tgt_ln));
676 676
677 677 avl_create(&iscsit_global.global_target_list,
678 678 iscsit_tgt_avl_compare, sizeof (iscsit_tgt_t),
679 679 offsetof(iscsit_tgt_t, target_global_ln));
680 680
681 681 list_create(&iscsit_global.global_deleted_target_list,
682 682 sizeof (iscsit_tgt_t),
683 683 offsetof(iscsit_tgt_t, target_global_deleted_ln));
684 684
685 685 avl_create(&iscsit_global.global_tpg_list,
686 686 iscsit_tpg_avl_compare, sizeof (iscsit_tpg_t),
687 687 offsetof(iscsit_tpg_t, tpg_global_ln));
688 688
689 689 avl_create(&iscsit_global.global_ini_list,
690 690 iscsit_ini_avl_compare, sizeof (iscsit_ini_t),
691 691 offsetof(iscsit_ini_t, ini_global_ln));
692 692
693 693 iscsit_global.global_tsih_pool = vmem_create("iscsit_tsih_pool",
694 694 (void *)1, ISCSI_MAX_TSIH, 1, NULL, NULL, NULL, 0,
695 695 VM_SLEEP | VMC_IDENTIFIER);
696 696
697 697 /*
698 698 * Setup STMF dbuf store. Our buffers are bound to a specific
699 699 * connection so we really can't let STMF cache buffers for us.
700 700 * Consequently we'll just allocate one global buffer store.
701 701 */
702 702 dbuf_store = stmf_alloc(STMF_STRUCT_DBUF_STORE, 0, 0);
703 703 if (dbuf_store == NULL) {
704 704 retval = IDM_STATUS_FAIL;
705 705 goto tear_down_and_return;
706 706 }
707 707 dbuf_store->ds_alloc_data_buf = iscsit_dbuf_alloc;
708 708 dbuf_store->ds_free_data_buf = iscsit_dbuf_free;
709 709 dbuf_store->ds_port_private = NULL;
710 710 iscsit_global.global_dbuf_store = dbuf_store;
711 711
712 712 /* Status PDU cache */
713 713 iscsit_status_pdu_cache = kmem_cache_create("iscsit_status_pdu_cache",
714 714 sizeof (idm_pdu_t) + sizeof (iscsi_scsi_rsp_hdr_t), 8,
715 715 &iscsit_status_pdu_constructor,
716 716 NULL, NULL, NULL, NULL, KM_SLEEP);
717 717
718 718 /* Default TPG and portal */
719 719 iscsit_global.global_default_tpg = iscsit_tpg_createdefault();
720 720 if (iscsit_global.global_default_tpg == NULL) {
721 721 retval = IDM_STATUS_FAIL;
722 722 goto tear_down_and_return;
723 723 }
724 724
725 725 /* initialize isns client */
726 726 (void) iscsit_isns_init(hostinfo);
727 727 did_iscsit_isns_init = B_TRUE;
728 728
729 729 /* Register port provider */
730 730 pp = stmf_alloc(STMF_STRUCT_PORT_PROVIDER, 0, 0);
731 731 if (pp == NULL) {
732 732 retval = IDM_STATUS_FAIL;
733 733 goto tear_down_and_return;
734 734 }
735 735
736 736 pp->pp_portif_rev = PORTIF_REV_1;
737 737 pp->pp_instance = 0;
738 738 pp->pp_name = ISCSIT_MODNAME;
739 739 pp->pp_cb = iscsit_pp_cb;
740 740
741 741 iscsit_global.global_pp = pp;
742 742
743 743
744 744 if (stmf_register_port_provider(pp) != STMF_SUCCESS) {
745 745 retval = IDM_STATUS_FAIL;
746 746 goto tear_down_and_return;
747 747 }
748 748
749 749 iscsit_global.global_dispatch_taskq = taskq_create("iscsit_dispatch",
750 750 1, minclsyspri, 16, 16, TASKQ_PREPOPULATE);
751 751
752 752 /* Scan staged PDUs, meaningful in MC/S situations */
753 753 iscsit_rxpdu_queue_monitor_start();
754 754
755 755 return (IDM_STATUS_SUCCESS);
756 756
757 757 tear_down_and_return:
758 758
759 759 if (iscsit_global.global_dispatch_taskq) {
760 760 taskq_destroy(iscsit_global.global_dispatch_taskq);
761 761 iscsit_global.global_dispatch_taskq = NULL;
762 762 }
763 763
764 764 if (did_iscsit_isns_init)
765 765 iscsit_isns_fini();
766 766
767 767 if (iscsit_global.global_default_tpg) {
768 768 iscsit_tpg_destroydefault(iscsit_global.global_default_tpg);
769 769 iscsit_global.global_default_tpg = NULL;
770 770 }
771 771
772 772 if (iscsit_global.global_pp)
773 773 iscsit_global.global_pp = NULL;
774 774
775 775 if (pp)
776 776 stmf_free(pp);
777 777
778 778 if (iscsit_status_pdu_cache) {
779 779 kmem_cache_destroy(iscsit_status_pdu_cache);
780 780 iscsit_status_pdu_cache = NULL;
781 781 }
782 782
783 783 if (iscsit_global.global_dbuf_store) {
784 784 stmf_free(iscsit_global.global_dbuf_store);
785 785 iscsit_global.global_dbuf_store = NULL;
786 786 }
787 787
788 788 if (iscsit_global.global_tsih_pool) {
789 789 vmem_destroy(iscsit_global.global_tsih_pool);
790 790 iscsit_global.global_tsih_pool = NULL;
791 791 }
792 792
793 793 avl_destroy(&iscsit_global.global_ini_list);
794 794 avl_destroy(&iscsit_global.global_tpg_list);
795 795 list_destroy(&iscsit_global.global_deleted_target_list);
796 796 avl_destroy(&iscsit_global.global_target_list);
797 797 avl_destroy(&iscsit_global.global_discovery_sessions);
798 798
799 799 idm_refcnt_destroy(&iscsit_global.global_refcnt);
800 800
801 801 return (retval);
802 802 }
803 803
804 804 /*
805 805 * iscsit_disable_svc
806 806 *
807 807 * clean up all existing connections and deregister targets from STMF
808 808 */
809 809 static void
810 810 iscsit_disable_svc(void)
811 811 {
812 812 iscsit_sess_t *sess;
813 813
814 814 ASSERT(iscsit_global.global_svc_state == ISE_DISABLING);
815 815
816 816 iscsit_rxpdu_queue_monitor_stop();
817 817
818 818 /* tear down discovery sessions */
819 819 for (sess = avl_first(&iscsit_global.global_discovery_sessions);
820 820 sess != NULL;
821 821 sess = AVL_NEXT(&iscsit_global.global_discovery_sessions, sess))
822 822 iscsit_sess_close(sess);
823 823
824 824 /*
825 825 * Passing NULL to iscsit_config_merge tells it to go to an empty
826 826 * config.
827 827 */
828 828 (void) iscsit_config_merge(NULL);
829 829
830 830 /*
831 831 * Wait until there are no more global references
832 832 */
833 833 idm_refcnt_wait_ref(&iscsit_global.global_refcnt);
834 834 idm_refcnt_destroy(&iscsit_global.global_refcnt);
835 835
836 836 /*
837 837 * Default TPG must be destroyed after global_refcnt is 0.
838 838 */
839 839 iscsit_tpg_destroydefault(iscsit_global.global_default_tpg);
840 840
841 841 avl_destroy(&iscsit_global.global_discovery_sessions);
842 842 list_destroy(&iscsit_global.global_deleted_target_list);
843 843 avl_destroy(&iscsit_global.global_target_list);
844 844 avl_destroy(&iscsit_global.global_tpg_list);
845 845 avl_destroy(&iscsit_global.global_ini_list);
846 846
847 847 taskq_destroy(iscsit_global.global_dispatch_taskq);
848 848
849 849 iscsit_isns_fini();
850 850
851 851 stmf_free(iscsit_global.global_dbuf_store);
852 852 iscsit_global.global_dbuf_store = NULL;
853 853
854 854 (void) stmf_deregister_port_provider(iscsit_global.global_pp);
855 855 stmf_free(iscsit_global.global_pp);
856 856 iscsit_global.global_pp = NULL;
857 857
858 858 kmem_cache_destroy(iscsit_status_pdu_cache);
859 859 iscsit_status_pdu_cache = NULL;
860 860
861 861 vmem_destroy(iscsit_global.global_tsih_pool);
862 862 iscsit_global.global_tsih_pool = NULL;
863 863 }
864 864
865 865 void
866 866 iscsit_global_hold()
867 867 {
868 868 /*
869 869 * To take out a global hold, we must either own the global
870 870 * state mutex or we must be running inside of an ioctl that
871 871 * has set the global state to ISE_BUSY, ISE_DISABLING, or
872 872 * ISE_ENABLING. We don't track the "owner" for these flags,
873 873 * so just checking if they are set is enough for now.
874 874 */
875 875 ASSERT((iscsit_global.global_svc_state == ISE_ENABLING) ||
876 876 (iscsit_global.global_svc_state == ISE_DISABLING) ||
877 877 (iscsit_global.global_svc_state == ISE_BUSY) ||
878 878 MUTEX_HELD(&iscsit_global.global_state_mutex));
879 879
880 880 idm_refcnt_hold(&iscsit_global.global_refcnt);
881 881 }
882 882
883 883 void
884 884 iscsit_global_rele()
885 885 {
886 886 idm_refcnt_rele(&iscsit_global.global_refcnt);
887 887 }
888 888
889 889 void
890 890 iscsit_global_wait_ref()
891 891 {
892 892 idm_refcnt_wait_ref(&iscsit_global.global_refcnt);
893 893 }
894 894
895 895 /*
896 896 * IDM callbacks
897 897 */
898 898
899 899 /*ARGSUSED*/
900 900 void
901 901 iscsit_rx_pdu(idm_conn_t *ic, idm_pdu_t *rx_pdu)
902 902 {
903 903 iscsit_conn_t *ict = ic->ic_handle;
904 904 switch (IDM_PDU_OPCODE(rx_pdu)) {
905 905 case ISCSI_OP_SCSI_CMD:
906 906 ASSERT(0); /* Shouldn't happen */
907 907 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
908 908 break;
909 909 case ISCSI_OP_SNACK_CMD:
910 910 /*
911 911 * We'll need to handle this when we support ERL1/2. For
912 912 * now we treat it as a protocol error.
913 913 */
914 914 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
915 915 idm_conn_event(ic, CE_TRANSPORT_FAIL, NULL);
916 916 break;
917 917 case ISCSI_OP_SCSI_TASK_MGT_MSG:
918 918 if (iscsit_check_cmdsn_and_queue(rx_pdu)) {
919 919 iscsit_set_cmdsn(ict, rx_pdu);
920 920 iscsit_op_scsi_task_mgmt(ict, rx_pdu);
921 921 }
922 922 break;
923 923 case ISCSI_OP_NOOP_OUT:
924 924 case ISCSI_OP_LOGIN_CMD:
925 925 case ISCSI_OP_TEXT_CMD:
926 926 case ISCSI_OP_LOGOUT_CMD:
927 927 /*
928 928 * If/when we switch to userland processing these PDU's
929 929 * will be handled by iscsitd.
930 930 */
931 931 iscsit_deferred_dispatch(rx_pdu);
932 932 break;
933 933 default:
934 934 /* Protocol error */
935 935 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
936 936 idm_conn_event(ic, CE_TRANSPORT_FAIL, NULL);
937 937 break;
938 938 }
939 939 }
940 940
941 941 /*ARGSUSED*/
942 942 void
943 943 iscsit_rx_pdu_error(idm_conn_t *ic, idm_pdu_t *rx_pdu, idm_status_t status)
944 944 {
945 945 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
946 946 }
947 947
948 948 void
949 949 iscsit_task_aborted(idm_task_t *idt, idm_status_t status)
950 950 {
951 951 iscsit_task_t *itask = idt->idt_private;
952 952
953 953 switch (status) {
954 954 case IDM_STATUS_SUSPENDED:
955 955 break;
956 956 case IDM_STATUS_ABORTED:
957 957 mutex_enter(&itask->it_mutex);
958 958 itask->it_aborted = B_TRUE;
959 959 /*
960 960 * We rely on the fact that STMF tracks outstanding
961 961 * buffer transfers and will free all of our buffers
962 962 * before freeing the task so we don't need to
963 963 * explicitly free the buffers from iscsit/idm
964 964 */
965 965 if (itask->it_stmf_abort) {
966 966 mutex_exit(&itask->it_mutex);
967 967 /*
968 968 * Task is no longer active
969 969 */
970 970 iscsit_task_done(itask);
971 971
972 972 /*
973 973 * STMF has already asked for this task to be aborted
974 974 *
975 975 * STMF specification is wrong... says to return
976 976 * STMF_ABORTED, the code actually looks for
977 977 * STMF_ABORT_SUCCESS.
978 978 */
979 979 stmf_task_lport_aborted(itask->it_stmf_task,
980 980 STMF_ABORT_SUCCESS, STMF_IOF_LPORT_DONE);
981 981 return;
982 982 } else {
983 983 mutex_exit(&itask->it_mutex);
984 984 /*
985 985 * Tell STMF to stop processing the task.
986 986 */
987 987 stmf_abort(STMF_QUEUE_TASK_ABORT, itask->it_stmf_task,
988 988 STMF_ABORTED, NULL);
989 989 return;
990 990 }
991 991 /*NOTREACHED*/
992 992 default:
993 993 ASSERT(0);
994 994 }
995 995 }
996 996
997 997 /*ARGSUSED*/
998 998 idm_status_t
999 999 iscsit_client_notify(idm_conn_t *ic, idm_client_notify_t icn,
1000 1000 uintptr_t data)
1001 1001 {
1002 1002 idm_status_t rc = IDM_STATUS_SUCCESS;
1003 1003
1004 1004 /*
1005 1005 * IDM client notifications will never occur at interrupt level
1006 1006 * since they are generated from the connection state machine which
1007 1007 * running on taskq threads.
1008 1008 *
1009 1009 */
1010 1010 switch (icn) {
1011 1011 case CN_CONNECT_ACCEPT:
1012 1012 rc = iscsit_conn_accept(ic); /* No data */
1013 1013 break;
1014 1014 case CN_FFP_ENABLED:
1015 1015 rc = iscsit_ffp_enabled(ic); /* No data */
1016 1016 break;
1017 1017 case CN_FFP_DISABLED:
1018 1018 /*
1019 1019 * Data indicates whether this was the result of an
1020 1020 * explicit logout request.
1021 1021 */
1022 1022 rc = iscsit_ffp_disabled(ic, (idm_ffp_disable_t)data);
1023 1023 break;
1024 1024 case CN_CONNECT_LOST:
1025 1025 rc = iscsit_conn_lost(ic);
1026 1026 break;
1027 1027 case CN_CONNECT_DESTROY:
1028 1028 rc = iscsit_conn_destroy(ic);
1029 1029 break;
1030 1030 case CN_LOGIN_FAIL:
1031 1031 /*
1032 1032 * Force the login state machine to completion
1033 1033 */
1034 1034 rc = iscsit_login_fail(ic);
1035 1035 break;
1036 1036 default:
1037 1037 rc = IDM_STATUS_REJECT;
1038 1038 break;
1039 1039 }
1040 1040
1041 1041 return (rc);
1042 1042 }
1043 1043
1044 1044 /*
1045 1045 * iscsit_update_statsn is invoked for all the PDUs which have the StatSN
1046 1046 * field in the header. The StatSN is incremented if the IDM_PDU_ADVANCE_STATSN
1047 1047 * flag is set in the pdu flags field. The StatSN is connection-wide and is
1048 1048 * protected by the mutex ict_statsn_mutex. For Data-In PDUs, if the flag
1049 1049 * IDM_TASK_PHASECOLLAPSE_REQ is set, the status (phase-collapse) is also filled
1050 1050 */
1051 1051 void
1052 1052 iscsit_update_statsn(idm_task_t *idm_task, idm_pdu_t *pdu)
1053 1053 {
1054 1054 iscsi_scsi_rsp_hdr_t *rsp = (iscsi_scsi_rsp_hdr_t *)pdu->isp_hdr;
1055 1055 iscsit_conn_t *ict = (iscsit_conn_t *)pdu->isp_ic->ic_handle;
1056 1056 iscsit_task_t *itask = NULL;
1057 1057 scsi_task_t *task = NULL;
1058 1058
1059 1059 mutex_enter(&ict->ict_statsn_mutex);
1060 1060 rsp->statsn = htonl(ict->ict_statsn);
1061 1061 if (pdu->isp_flags & IDM_PDU_ADVANCE_STATSN)
1062 1062 ict->ict_statsn++;
1063 1063 mutex_exit(&ict->ict_statsn_mutex);
1064 1064
1065 1065 /*
1066 1066 * The last SCSI Data PDU passed for a command may also contain the
1067 1067 * status if the status indicates termination with no expections, i.e.
1068 1068 * no sense data or response involved. If the command completes with
1069 1069 * an error, then the response and sense data will be sent in a
1070 1070 * separate iSCSI Response PDU.
1071 1071 */
1072 1072 if ((idm_task) && (idm_task->idt_flags & IDM_TASK_PHASECOLLAPSE_REQ)) {
1073 1073 itask = idm_task->idt_private;
1074 1074 task = itask->it_stmf_task;
1075 1075
1076 1076 rsp->cmd_status = task->task_scsi_status;
1077 1077 rsp->flags |= ISCSI_FLAG_DATA_STATUS;
1078 1078 if (task->task_status_ctrl & TASK_SCTRL_OVER) {
1079 1079 rsp->flags |= ISCSI_FLAG_CMD_OVERFLOW;
1080 1080 } else if (task->task_status_ctrl & TASK_SCTRL_UNDER) {
1081 1081 rsp->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
1082 1082 }
1083 1083 rsp->residual_count = htonl(task->task_resid);
1084 1084
1085 1085 /*
1086 1086 * Removing the task from the session task list
1087 1087 * just before the status is sent in the last
1088 1088 * Data PDU transfer
1089 1089 */
1090 1090 iscsit_task_done(itask);
1091 1091 }
1092 1092 }
1093 1093
1094 1094 void
1095 1095 iscsit_build_hdr(idm_task_t *idm_task, idm_pdu_t *pdu, uint8_t opcode)
1096 1096 {
1097 1097 iscsit_task_t *itask = idm_task->idt_private;
1098 1098 iscsi_data_rsp_hdr_t *dh = (iscsi_data_rsp_hdr_t *)pdu->isp_hdr;
1099 1099
1100 1100 /*
1101 1101 * We acquired iscsit_sess_t.ist_sn_mutex in iscsit_xfer_scsi_data
1102 1102 */
1103 1103 ASSERT(MUTEX_HELD(&itask->it_ict->ict_sess->ist_sn_mutex));
1104 1104 /*
1105 1105 * On incoming data, the target transfer tag and Lun is only
1106 1106 * provided by the target if the A bit is set, Since the target
1107 1107 * does not currently support Error Recovery Level 1, the A
1108 1108 * bit is never set.
1109 1109 */
1110 1110 dh->opcode = opcode;
1111 1111 dh->itt = itask->it_itt;
1112 1112 dh->ttt = ((opcode & ISCSI_OPCODE_MASK) == ISCSI_OP_SCSI_DATA_RSP) ?
1113 1113 ISCSI_RSVD_TASK_TAG : itask->it_ttt;
1114 1114
1115 1115 dh->expcmdsn = htonl(itask->it_ict->ict_sess->ist_expcmdsn);
1116 1116 dh->maxcmdsn = htonl(itask->it_ict->ict_sess->ist_maxcmdsn);
1117 1117
1118 1118 /*
1119 1119 * IDM must set:
1120 1120 *
1121 1121 * data.flags and rtt.flags
1122 1122 * data.dlength
1123 1123 * data.datasn
1124 1124 * data.offset
1125 1125 * statsn, residual_count and cmd_status (for phase collapse)
1126 1126 * rtt.rttsn
1127 1127 * rtt.data_offset
1128 1128 * rtt.data_length
1129 1129 */
1130 1130 }
1131 1131
1132 1132 void
1133 1133 iscsit_keepalive(idm_conn_t *ic)
1134 1134 {
1135 1135 idm_pdu_t *nop_in_pdu;
1136 1136 iscsi_nop_in_hdr_t *nop_in;
1137 1137 iscsit_conn_t *ict = ic->ic_handle;
1138 1138
1139 1139 /*
1140 1140 * IDM noticed the connection has been idle for too long so it's
1141 1141 * time to provoke some activity. Build and transmit an iSCSI
1142 1142 * nop-in PDU -- when the initiator responds it will be counted
1143 1143 * as "activity" and keep the connection alive.
1144 1144 *
1145 1145 * We don't actually care about the response here at the iscsit level
1146 1146 * so we will just throw it away without looking at it when it arrives.
1147 1147 */
1148 1148 nop_in_pdu = idm_pdu_alloc(sizeof (*nop_in), 0);
1149 1149 idm_pdu_init(nop_in_pdu, ic, NULL, NULL);
1150 1150 nop_in = (iscsi_nop_in_hdr_t *)nop_in_pdu->isp_hdr;
1151 1151 bzero(nop_in, sizeof (*nop_in));
1152 1152 nop_in->opcode = ISCSI_OP_NOOP_IN;
1153 1153 nop_in->flags = ISCSI_FLAG_FINAL;
1154 1154 nop_in->itt = ISCSI_RSVD_TASK_TAG;
1155 1155 /*
1156 1156 * When the target sends a NOP-In as a Ping, the target transfer tag
1157 1157 * is set to a valid (not reserved) value and the initiator task tag
1158 1158 * is set to ISCSI_RSVD_TASK_TAG (0xffffffff). In this case the StatSN
1159 1159 * will always contain the next sequence number but the StatSN for the
1160 1160 * connection is not advanced after this PDU is sent.
1161 1161 */
1162 1162 nop_in_pdu->isp_flags |= IDM_PDU_SET_STATSN;
1163 1163 /*
1164 1164 * This works because we don't currently allocate ttt's anywhere else
1165 1165 * in iscsit so as long as we stay out of IDM's range we are safe.
1166 1166 * If we need to allocate ttt's for other PDU's in the future this will
1167 1167 * need to be improved.
1168 1168 */
1169 1169 mutex_enter(&ict->ict_mutex);
1170 1170 nop_in->ttt = ict->ict_keepalive_ttt;
1171 1171 ict->ict_keepalive_ttt++;
1172 1172 if (ict->ict_keepalive_ttt == ISCSI_RSVD_TASK_TAG)
1173 1173 ict->ict_keepalive_ttt = IDM_TASKIDS_MAX;
1174 1174 mutex_exit(&ict->ict_mutex);
1175 1175
1176 1176 iscsit_pdu_tx(nop_in_pdu);
1177 1177 }
1178 1178
1179 1179 static idm_status_t
1180 1180 iscsit_conn_accept(idm_conn_t *ic)
1181 1181 {
1182 1182 iscsit_conn_t *ict;
1183 1183
1184 1184 /*
1185 1185 * We need to get a global hold here to ensure that the service
1186 1186 * doesn't get shutdown prior to establishing a session. This
1187 1187 * gets released in iscsit_conn_destroy().
1188 1188 */
1189 1189 mutex_enter(&iscsit_global.global_state_mutex);
1190 1190 if (iscsit_global.global_svc_state != ISE_ENABLED) {
1191 1191 mutex_exit(&iscsit_global.global_state_mutex);
1192 1192 return (IDM_STATUS_FAIL);
1193 1193 }
1194 1194 iscsit_global_hold();
1195 1195 mutex_exit(&iscsit_global.global_state_mutex);
1196 1196
1197 1197 /*
1198 1198 * Allocate an associated iscsit structure to represent this
1199 1199 * connection. We shouldn't really create a session until we
1200 1200 * get the first login PDU.
1201 1201 */
1202 1202 ict = kmem_zalloc(sizeof (*ict), KM_SLEEP);
1203 1203
1204 1204 ict->ict_ic = ic;
1205 1205 ict->ict_statsn = 1;
1206 1206 ict->ict_keepalive_ttt = IDM_TASKIDS_MAX; /* Avoid IDM TT range */
1207 1207 ic->ic_handle = ict;
1208 1208 mutex_init(&ict->ict_mutex, NULL, MUTEX_DRIVER, NULL);
1209 1209 mutex_init(&ict->ict_statsn_mutex, NULL, MUTEX_DRIVER, NULL);
1210 1210 idm_refcnt_init(&ict->ict_refcnt, ict);
1211 1211
1212 1212 /*
1213 1213 * Initialize login state machine
1214 1214 */
1215 1215 if (iscsit_login_sm_init(ict) != IDM_STATUS_SUCCESS) {
1216 1216 iscsit_global_rele();
1217 1217 /*
1218 1218 * Cleanup the ict after idm notifies us about this failure
1219 1219 */
1220 1220 return (IDM_STATUS_FAIL);
1221 1221 }
1222 1222
1223 1223 return (IDM_STATUS_SUCCESS);
1224 1224 }
1225 1225
1226 1226 idm_status_t
1227 1227 iscsit_conn_reinstate(iscsit_conn_t *reinstate_ict, iscsit_conn_t *new_ict)
1228 1228 {
1229 1229 idm_status_t result;
1230 1230
1231 1231 /*
1232 1232 * Note in new connection state that this connection is
1233 1233 * reinstating an existing connection.
1234 1234 */
1235 1235 new_ict->ict_reinstating = B_TRUE;
1236 1236 new_ict->ict_reinstate_conn = reinstate_ict;
1237 1237 new_ict->ict_statsn = reinstate_ict->ict_statsn;
1238 1238
1239 1239 /*
1240 1240 * Now generate connection state machine event to existing connection
1241 1241 * so that it starts the cleanup process.
1242 1242 */
1243 1243 result = idm_conn_reinstate_event(reinstate_ict->ict_ic,
1244 1244 new_ict->ict_ic);
1245 1245
1246 1246 return (result);
1247 1247 }
1248 1248
1249 1249 void
1250 1250 iscsit_conn_hold(iscsit_conn_t *ict)
1251 1251 {
1252 1252 idm_refcnt_hold(&ict->ict_refcnt);
1253 1253 }
1254 1254
1255 1255 void
1256 1256 iscsit_conn_rele(iscsit_conn_t *ict)
1257 1257 {
1258 1258 idm_refcnt_rele(&ict->ict_refcnt);
1259 1259 }
1260 1260
1261 1261 void
1262 1262 iscsit_conn_dispatch_hold(iscsit_conn_t *ict)
1263 1263 {
1264 1264 idm_refcnt_hold(&ict->ict_dispatch_refcnt);
1265 1265 }
1266 1266
1267 1267 void
1268 1268 iscsit_conn_dispatch_rele(iscsit_conn_t *ict)
1269 1269 {
1270 1270 idm_refcnt_rele(&ict->ict_dispatch_refcnt);
1271 1271 }
1272 1272
1273 1273 static idm_status_t
1274 1274 iscsit_login_fail(idm_conn_t *ic)
1275 1275 {
1276 1276 iscsit_conn_t *ict = ic->ic_handle;
1277 1277
1278 1278 /* Generate login state machine event */
1279 1279 iscsit_login_sm_event(ict, ILE_LOGIN_CONN_ERROR, NULL);
1280 1280
1281 1281 return (IDM_STATUS_SUCCESS);
1282 1282 }
1283 1283
1284 1284 static idm_status_t
1285 1285 iscsit_ffp_enabled(idm_conn_t *ic)
1286 1286 {
1287 1287 iscsit_conn_t *ict = ic->ic_handle;
1288 1288
1289 1289 /* Generate session state machine event */
1290 1290 iscsit_sess_sm_event(ict->ict_sess, SE_CONN_LOGGED_IN, ict);
1291 1291
1292 1292 return (IDM_STATUS_SUCCESS);
1293 1293 }
1294 1294
1295 1295 static idm_status_t
1296 1296 iscsit_ffp_disabled(idm_conn_t *ic, idm_ffp_disable_t disable_class)
1297 1297 {
1298 1298 iscsit_conn_t *ict = ic->ic_handle;
1299 1299
1300 1300 /* Generate session state machine event */
1301 1301 switch (disable_class) {
1302 1302 case FD_CONN_FAIL:
1303 1303 iscsit_sess_sm_event(ict->ict_sess, SE_CONN_FFP_FAIL, ict);
1304 1304 break;
1305 1305 case FD_CONN_LOGOUT:
1306 1306 iscsit_sess_sm_event(ict->ict_sess, SE_CONN_FFP_DISABLE, ict);
1307 1307 break;
1308 1308 case FD_SESS_LOGOUT:
1309 1309 iscsit_sess_sm_event(ict->ict_sess, SE_SESSION_CLOSE, ict);
1310 1310 break;
1311 1311 default:
1312 1312 ASSERT(0);
1313 1313 }
1314 1314
1315 1315 return (IDM_STATUS_SUCCESS);
1316 1316 }
1317 1317
1318 1318 static idm_status_t
1319 1319 iscsit_conn_lost(idm_conn_t *ic)
1320 1320 {
1321 1321 iscsit_conn_t *ict = ic->ic_handle;
1322 1322 iscsit_sess_t *ist = ict->ict_sess;
1323 1323 iscsit_cbuf_t *cbuf;
1324 1324 idm_pdu_t *rx_pdu;
1325 1325 int i;
1326 1326
1327 1327 mutex_enter(&ict->ict_mutex);
1328 1328 ict->ict_lost = B_TRUE;
1329 1329 mutex_exit(&ict->ict_mutex);
1330 1330 /*
1331 1331 * scrub the staging queue for all PDUs on this connection
1332 1332 */
1333 1333 if (ist != NULL) {
1334 1334 mutex_enter(&ist->ist_sn_mutex);
1335 1335 for (cbuf = ist->ist_rxpdu_queue, i = 0;
1336 1336 ((cbuf->cb_num_elems > 0) && (i < ISCSIT_RXPDU_QUEUE_LEN));
1337 1337 i++) {
1338 1338 if (((rx_pdu = cbuf->cb_buffer[i]) != NULL) &&
1339 1339 (rx_pdu->isp_ic == ic)) {
1340 1340 /* conn is lost, drop the pdu */
1341 1341 DTRACE_PROBE3(scrubbing__staging__queue,
1342 1342 iscsit_sess_t *, ist, idm_conn_t *, ic,
1343 1343 idm_pdu_t *, rx_pdu);
1344 1344 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
1345 1345 cbuf->cb_buffer[i] = NULL;
1346 1346 cbuf->cb_num_elems--;
1347 1347 iscsit_conn_dispatch_rele(ict);
1348 1348 }
1349 1349 }
1350 1350 mutex_exit(&ist->ist_sn_mutex);
1351 1351 }
1352 1352 /*
1353 1353 * Make sure there aren't any PDU's transitioning from the receive
1354 1354 * handler to the dispatch taskq.
1355 1355 */
1356 1356 idm_refcnt_wait_ref(&ict->ict_dispatch_refcnt);
1357 1357
1358 1358 return (IDM_STATUS_SUCCESS);
1359 1359 }
1360 1360
1361 1361 static idm_status_t
1362 1362 iscsit_conn_destroy(idm_conn_t *ic)
1363 1363 {
1364 1364 iscsit_conn_t *ict = ic->ic_handle;
1365 1365
1366 1366 mutex_enter(&ict->ict_mutex);
1367 1367 ict->ict_destroyed = B_TRUE;
1368 1368 mutex_exit(&ict->ict_mutex);
1369 1369
1370 1370 /* Generate session state machine event */
1371 1371 if (ict->ict_sess != NULL) {
1372 1372 /*
1373 1373 * Session state machine will call iscsit_conn_destroy_done()
1374 1374 * when it has removed references to this connection.
1375 1375 */
1376 1376 iscsit_sess_sm_event(ict->ict_sess, SE_CONN_FAIL, ict);
1377 1377 }
1378 1378
1379 1379 idm_refcnt_wait_ref(&ict->ict_refcnt);
1380 1380 /*
1381 1381 * The session state machine does not need to post
1382 1382 * events to IDM any longer, so it is safe to set
1383 1383 * the idm connection reference to NULL
1384 1384 */
1385 1385 ict->ict_ic = NULL;
1386 1386
1387 1387 /* Reap the login state machine */
1388 1388 iscsit_login_sm_fini(ict);
1389 1389
1390 1390 /* Clean up any text command remnants */
1391 1391 iscsit_text_cmd_fini(ict);
1392 1392
1393 1393 mutex_destroy(&ict->ict_mutex);
1394 1394 idm_refcnt_destroy(&ict->ict_refcnt);
1395 1395 kmem_free(ict, sizeof (*ict));
1396 1396
1397 1397 iscsit_global_rele();
1398 1398
1399 1399 return (IDM_STATUS_SUCCESS);
1400 1400 }
1401 1401
1402 1402 void
1403 1403 iscsit_conn_logout(iscsit_conn_t *ict)
1404 1404 {
1405 1405 /*
1406 1406 * If the iscsi connection is active, then
1407 1407 * logout the IDM connection by sending a
1408 1408 * CE_LOGOUT_SESSION_SUCCESS, else, no action
1409 1409 * needs to be taken because the connection
1410 1410 * is already in the teardown process.
1411 1411 */
1412 1412 mutex_enter(&ict->ict_mutex);
1413 1413 if (ict->ict_lost == B_FALSE && ict->ict_destroyed == B_FALSE) {
1414 1414 idm_conn_event(ict->ict_ic, CE_LOGOUT_SESSION_SUCCESS, NULL);
1415 1415 }
1416 1416 mutex_exit(&ict->ict_mutex);
1417 1417 }
1418 1418
1419 1419 /*
1420 1420 * STMF-related functions
1421 1421 *
1422 1422 * iSCSI to STMF mapping
1423 1423 *
1424 1424 * Session == ?
1425 1425 * Connection == bound to local port but not itself a local port
1426 1426 * Target
1427 1427 * Target portal (group?) == local port (really but we're not going to do this)
1428 1428 * iscsit needs to map connections to local ports (whatever we decide
1429 1429 * they are)
1430 1430 * Target == ?
1431 1431 */
1432 1432
1433 1433 /*ARGSUSED*/
1434 1434 static stmf_data_buf_t *
1435 1435 iscsit_dbuf_alloc(scsi_task_t *task, uint32_t size, uint32_t *pminsize,
1436 1436 uint32_t flags)
1437 1437 {
1438 1438 iscsit_task_t *itask = task->task_port_private;
1439 1439 idm_buf_t *idm_buffer;
1440 1440 iscsit_buf_t *ibuf;
1441 1441 stmf_data_buf_t *result;
1442 1442 uint32_t bsize;
1443 1443
1444 1444 /*
1445 1445 * If the requested size is larger than MaxBurstLength and the
1446 1446 * given pminsize is also larger than MaxBurstLength, then the
1447 1447 * allocation fails (dbuf = NULL) and pminsize is modified to
1448 1448 * be equal to MaxBurstLength. stmf/sbd then should re-invoke
1449 1449 * this function with the corrected values for transfer.
1450 1450 */
1451 1451 ASSERT(pminsize);
1452 1452 if (size <= itask->it_ict->ict_op.op_max_burst_length) {
1453 1453 bsize = size;
1454 1454 } else if (*pminsize <= itask->it_ict->ict_op.op_max_burst_length) {
1455 1455 bsize = itask->it_ict->ict_op.op_max_burst_length;
1456 1456 } else {
1457 1457 *pminsize = itask->it_ict->ict_op.op_max_burst_length;
1458 1458 return (NULL);
1459 1459 }
1460 1460
1461 1461 /* Alloc buffer */
1462 1462 idm_buffer = idm_buf_alloc(itask->it_ict->ict_ic, NULL, bsize);
1463 1463 if (idm_buffer != NULL) {
1464 1464 result = stmf_alloc(STMF_STRUCT_DATA_BUF,
1465 1465 sizeof (iscsit_buf_t), 0);
1466 1466 if (result != NULL) {
1467 1467 /* Fill in stmf_data_buf_t */
1468 1468 ibuf = result->db_port_private;
1469 1469 ibuf->ibuf_idm_buf = idm_buffer;
1470 1470 ibuf->ibuf_stmf_buf = result;
1471 1471 ibuf->ibuf_is_immed = B_FALSE;
1472 1472 result->db_flags = DB_DONT_CACHE;
1473 1473 result->db_buf_size = bsize;
1474 1474 result->db_data_size = bsize;
1475 1475 result->db_sglist_length = 1;
1476 1476 result->db_sglist[0].seg_addr = idm_buffer->idb_buf;
1477 1477 result->db_sglist[0].seg_length =
1478 1478 idm_buffer->idb_buflen;
1479 1479 return (result);
1480 1480 }
1481 1481
1482 1482 /* Couldn't get the stmf_data_buf_t so free the buffer */
1483 1483 idm_buf_free(idm_buffer);
1484 1484 }
1485 1485
1486 1486 return (NULL);
1487 1487 }
1488 1488
1489 1489 /*ARGSUSED*/
1490 1490 static void
1491 1491 iscsit_dbuf_free(stmf_dbuf_store_t *ds, stmf_data_buf_t *dbuf)
1492 1492 {
1493 1493 iscsit_buf_t *ibuf = dbuf->db_port_private;
1494 1494
1495 1495 if (ibuf->ibuf_is_immed) {
1496 1496 /*
1497 1497 * The iscsit_buf_t structure itself will be freed with its
1498 1498 * associated task. Here we just need to free the PDU that
1499 1499 * held the immediate data.
1500 1500 */
1501 1501 idm_pdu_complete(ibuf->ibuf_immed_data_pdu, IDM_STATUS_SUCCESS);
1502 1502 ibuf->ibuf_immed_data_pdu = 0;
1503 1503 } else {
1504 1504 idm_buf_free(ibuf->ibuf_idm_buf);
1505 1505 stmf_free(dbuf);
1506 1506 }
1507 1507 }
1508 1508
1509 1509 /*ARGSUSED*/
1510 1510 stmf_status_t
1511 1511 iscsit_xfer_scsi_data(scsi_task_t *task, stmf_data_buf_t *dbuf,
1512 1512 uint32_t ioflags)
1513 1513 {
1514 1514 iscsit_task_t *iscsit_task = task->task_port_private;
1515 1515 iscsit_sess_t *ict_sess = iscsit_task->it_ict->ict_sess;
1516 1516 iscsit_buf_t *ibuf = dbuf->db_port_private;
1517 1517 int idm_rc;
1518 1518
1519 1519 /*
1520 1520 * If we are aborting then we can ignore this request
1521 1521 */
1522 1522 if (iscsit_task->it_stmf_abort) {
1523 1523 return (STMF_SUCCESS);
1524 1524 }
1525 1525
1526 1526 /*
1527 1527 * If it's not immediate data then start the transfer
1528 1528 */
1529 1529 ASSERT(ibuf->ibuf_is_immed == B_FALSE);
1530 1530 if (dbuf->db_flags & DB_DIRECTION_TO_RPORT) {
1531 1531 /*
1532 1532 * The DB_SEND_STATUS_GOOD flag in the STMF data buffer allows
1533 1533 * the port provider to phase-collapse, i.e. send the status
1534 1534 * along with the final data PDU for the command. The port
1535 1535 * provider passes this request to the transport layer by
1536 1536 * setting a flag IDM_TASK_PHASECOLLAPSE_REQ in the task.
1537 1537 */
1538 1538 if (dbuf->db_flags & DB_SEND_STATUS_GOOD)
1539 1539 iscsit_task->it_idm_task->idt_flags |=
1540 1540 IDM_TASK_PHASECOLLAPSE_REQ;
1541 1541 /*
1542 1542 * IDM will call iscsit_build_hdr so lock now to serialize
1543 1543 * access to the SN values. We need to lock here to enforce
1544 1544 * lock ordering
1545 1545 */
1546 1546 mutex_enter(&ict_sess->ist_sn_mutex);
1547 1547 idm_rc = idm_buf_tx_to_ini(iscsit_task->it_idm_task,
1548 1548 ibuf->ibuf_idm_buf, dbuf->db_relative_offset,
1549 1549 dbuf->db_data_size, &iscsit_buf_xfer_cb, dbuf);
1550 1550 mutex_exit(&ict_sess->ist_sn_mutex);
1551 1551
1552 1552 return (iscsit_idm_to_stmf(idm_rc));
1553 1553 } else if (dbuf->db_flags & DB_DIRECTION_FROM_RPORT) {
1554 1554 /* Grab the SN lock (see comment above) */
1555 1555 mutex_enter(&ict_sess->ist_sn_mutex);
1556 1556 idm_rc = idm_buf_rx_from_ini(iscsit_task->it_idm_task,
1557 1557 ibuf->ibuf_idm_buf, dbuf->db_relative_offset,
1558 1558 dbuf->db_data_size, &iscsit_buf_xfer_cb, dbuf);
1559 1559 mutex_exit(&ict_sess->ist_sn_mutex);
1560 1560
1561 1561 return (iscsit_idm_to_stmf(idm_rc));
1562 1562 }
1563 1563
1564 1564 /* What are we supposed to do if there is no direction? */
1565 1565 return (STMF_INVALID_ARG);
1566 1566 }
1567 1567
1568 1568 static void
1569 1569 iscsit_buf_xfer_cb(idm_buf_t *idb, idm_status_t status)
1570 1570 {
1571 1571 iscsit_task_t *itask = idb->idb_task_binding->idt_private;
1572 1572 stmf_data_buf_t *dbuf = idb->idb_cb_arg;
1573 1573
1574 1574 dbuf->db_xfer_status = iscsit_idm_to_stmf(status);
1575 1575
1576 1576 /*
1577 1577 * If the task has been aborted then we don't need to call STMF
1578 1578 */
1579 1579 if (itask->it_stmf_abort) {
1580 1580 return;
1581 1581 }
1582 1582
1583 1583 /*
1584 1584 * For ISCSI over TCP (not iSER), the last SCSI Data PDU passed
1585 1585 * for a successful command contains the status as requested by
1586 1586 * by COMSTAR (via the DB_SEND_STATUS_GOOD flag). But the iSER
1587 1587 * transport does not support phase-collapse. So pretend we are
1588 1588 * COMSTAR and send the status in a separate PDU now.
1589 1589 */
1590 1590 if (idb->idb_task_binding->idt_flags & IDM_TASK_PHASECOLLAPSE_SUCCESS) {
1591 1591 /*
1592 1592 * Mark task complete and notify COMSTAR
1593 1593 * that the status has been sent.
1594 1594 */
1595 1595 itask->it_idm_task->idt_state = TASK_COMPLETE;
1596 1596 stmf_send_status_done(itask->it_stmf_task,
1597 1597 iscsit_idm_to_stmf(status), STMF_IOF_LPORT_DONE);
1598 1598 } else if ((dbuf->db_flags & DB_SEND_STATUS_GOOD) &&
1599 1599 status == IDM_STATUS_SUCCESS) {
1600 1600
1601 1601 /*
1602 1602 * The iscsi target port provider - for iSER, emulates the
1603 1603 * DB_SEND_STATUS_GOOD optimization if requested by STMF;
1604 1604 * it sends the status in a separate PDU after the data
1605 1605 * transfer. In this case the port provider should first
1606 1606 * call stmf_data_xfer_done() to mark the transfer complete
1607 1607 * and then send the status. Although STMF will free the
1608 1608 * buffer at the time the task is freed, even if the transfer
1609 1609 * is not marked complete, this behavior makes statistics
1610 1610 * gathering and task state tracking more difficult than it
1611 1611 * needs to be.
1612 1612 */
1613 1613 stmf_data_xfer_done(itask->it_stmf_task, dbuf, 0);
1614 1614 if (iscsit_send_scsi_status(itask->it_stmf_task, 0)
1615 1615 != STMF_SUCCESS) {
1616 1616 stmf_send_status_done(itask->it_stmf_task,
1617 1617 STMF_FAILURE, STMF_IOF_LPORT_DONE);
1618 1618 }
1619 1619 } else {
1620 1620 stmf_data_xfer_done(itask->it_stmf_task, dbuf, 0);
1621 1621 /* don't touch dbuf after stmf_data_xfer_done */
1622 1622 }
1623 1623 }
1624 1624
1625 1625
1626 1626 /*ARGSUSED*/
1627 1627 stmf_status_t
1628 1628 iscsit_send_scsi_status(scsi_task_t *task, uint32_t ioflags)
1629 1629 {
1630 1630 iscsit_task_t *itask = task->task_port_private;
1631 1631 iscsi_scsi_rsp_hdr_t *rsp;
1632 1632 idm_pdu_t *pdu;
1633 1633 int resp_datalen;
1634 1634
1635 1635 /*
1636 1636 * If this task is aborted then we don't need to respond.
1637 1637 */
1638 1638 if (itask->it_stmf_abort) {
1639 1639 return (STMF_SUCCESS);
1640 1640 }
1641 1641
1642 1642 /*
1643 1643 * If this is a task management status, handle it elsewhere.
1644 1644 */
1645 1645 if (task->task_mgmt_function != TM_NONE) {
1646 1646 /*
1647 1647 * Don't wait for the PDU completion to tell STMF
1648 1648 * the task is done -- it doesn't really matter and
1649 1649 * it makes life complicated if STMF later asks us to
1650 1650 * abort the request and we don't know whether the
1651 1651 * status has been sent or not.
1652 1652 */
1653 1653 itask->it_tm_responded = B_TRUE;
1654 1654 iscsit_send_task_mgmt_resp(itask->it_tm_pdu,
1655 1655 (task->task_completion_status == STMF_SUCCESS) ?
1656 1656 SCSI_TCP_TM_RESP_COMPLETE : SCSI_TCP_TM_RESP_FUNC_NOT_SUPP);
1657 1657 stmf_send_status_done(task, STMF_SUCCESS,
1658 1658 STMF_IOF_LPORT_DONE);
1659 1659 return (STMF_SUCCESS);
1660 1660 }
1661 1661
1662 1662 /*
1663 1663 * Remove the task from the session task list
1664 1664 */
1665 1665 iscsit_task_done(itask);
1666 1666
1667 1667 /*
1668 1668 * Send status
1669 1669 */
1670 1670 mutex_enter(&itask->it_idm_task->idt_mutex);
1671 1671 if ((itask->it_idm_task->idt_state == TASK_ACTIVE) &&
1672 1672 (task->task_completion_status == STMF_SUCCESS) &&
1673 1673 (task->task_sense_length == 0) &&
1674 1674 (task->task_resid == 0)) {
1675 1675 itask->it_idm_task->idt_state = TASK_COMPLETE;
1676 1676 /* PDU callback releases task hold */
1677 1677 idm_task_hold(itask->it_idm_task);
1678 1678 mutex_exit(&itask->it_idm_task->idt_mutex);
1679 1679 /*
1680 1680 * Fast path. Cached status PDU's are already
1681 1681 * initialized. We just need to fill in
1682 1682 * connection and task information. StatSN is
1683 1683 * incremented by 1 for every status sent a
1684 1684 * connection.
1685 1685 */
1686 1686 pdu = kmem_cache_alloc(iscsit_status_pdu_cache, KM_SLEEP);
1687 1687 pdu->isp_ic = itask->it_ict->ict_ic;
1688 1688 pdu->isp_private = itask;
1689 1689 pdu->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
1690 1690
1691 1691 rsp = (iscsi_scsi_rsp_hdr_t *)pdu->isp_hdr;
1692 1692 rsp->itt = itask->it_itt;
1693 1693 /*
1694 1694 * ExpDataSN is the number of R2T and Data-In (read)
1695 1695 * PDUs the target has sent for the SCSI command.
1696 1696 *
1697 1697 * Since there is no support for bidirectional transfer
1698 1698 * yet, either idt_exp_datasn or idt_exp_rttsn, but not
1699 1699 * both is valid at any time
1700 1700 */
1701 1701 rsp->expdatasn = (itask->it_idm_task->idt_exp_datasn != 0) ?
1702 1702 htonl(itask->it_idm_task->idt_exp_datasn):
1703 1703 htonl(itask->it_idm_task->idt_exp_rttsn);
1704 1704 rsp->cmd_status = task->task_scsi_status;
1705 1705 iscsit_pdu_tx(pdu);
1706 1706 return (STMF_SUCCESS);
1707 1707 } else {
1708 1708 if (itask->it_idm_task->idt_state != TASK_ACTIVE) {
1709 1709 mutex_exit(&itask->it_idm_task->idt_mutex);
1710 1710 return (STMF_FAILURE);
1711 1711 }
1712 1712 itask->it_idm_task->idt_state = TASK_COMPLETE;
1713 1713 /* PDU callback releases task hold */
1714 1714 idm_task_hold(itask->it_idm_task);
1715 1715 mutex_exit(&itask->it_idm_task->idt_mutex);
1716 1716
1717 1717 resp_datalen = (task->task_sense_length == 0) ? 0 :
1718 1718 (task->task_sense_length + sizeof (uint16_t));
1719 1719
1720 1720 pdu = idm_pdu_alloc(sizeof (iscsi_hdr_t), resp_datalen);
1721 1721 idm_pdu_init(pdu, itask->it_ict->ict_ic, itask,
1722 1722 iscsit_send_status_done);
1723 1723 pdu->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
1724 1724
1725 1725 rsp = (iscsi_scsi_rsp_hdr_t *)pdu->isp_hdr;
1726 1726 bzero(rsp, sizeof (*rsp));
1727 1727 rsp->opcode = ISCSI_OP_SCSI_RSP;
1728 1728
1729 1729 rsp->flags = ISCSI_FLAG_FINAL;
1730 1730 if (task->task_status_ctrl & TASK_SCTRL_OVER) {
1731 1731 rsp->flags |= ISCSI_FLAG_CMD_OVERFLOW;
1732 1732 } else if (task->task_status_ctrl & TASK_SCTRL_UNDER) {
1733 1733 rsp->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
1734 1734 }
1735 1735
1736 1736 rsp->bi_residual_count = 0;
1737 1737 rsp->residual_count = htonl(task->task_resid);
1738 1738 rsp->itt = itask->it_itt;
1739 1739 rsp->response = ISCSI_STATUS_CMD_COMPLETED;
1740 1740 rsp->expdatasn = (itask->it_idm_task->idt_exp_datasn != 0) ?
1741 1741 htonl(itask->it_idm_task->idt_exp_datasn):
1742 1742 htonl(itask->it_idm_task->idt_exp_rttsn);
1743 1743 rsp->cmd_status = task->task_scsi_status;
1744 1744 if (task->task_sense_length != 0) {
1745 1745 /*
1746 1746 * Add a byte to provide the sense length in
1747 1747 * the response
1748 1748 */
1749 1749 *(uint16_t *)((void *)pdu->isp_data) =
1750 1750 htons(task->task_sense_length);
1751 1751 bcopy(task->task_sense_data,
1752 1752 (uint8_t *)pdu->isp_data +
1753 1753 sizeof (uint16_t),
1754 1754 task->task_sense_length);
1755 1755 hton24(rsp->dlength, resp_datalen);
1756 1756 }
1757 1757
1758 1758 DTRACE_PROBE5(iscsi__scsi__response,
1759 1759 iscsit_conn_t *, itask->it_ict,
1760 1760 uint8_t, rsp->response,
1761 1761 uint8_t, rsp->cmd_status,
1762 1762 idm_pdu_t *, pdu,
1763 1763 scsi_task_t *, task);
1764 1764
1765 1765 iscsit_pdu_tx(pdu);
1766 1766
1767 1767 return (STMF_SUCCESS);
1768 1768 }
1769 1769 }
1770 1770
1771 1771 /*ARGSUSED*/
1772 1772 static void
1773 1773 iscsit_send_good_status_done(idm_pdu_t *pdu, idm_status_t status)
1774 1774 {
1775 1775 iscsit_task_t *itask;
1776 1776 boolean_t aborted;
1777 1777
1778 1778 itask = pdu->isp_private;
1779 1779 aborted = itask->it_stmf_abort;
1780 1780
1781 1781 /*
1782 1782 * After releasing the hold the task may be freed at any time so
1783 1783 * don't touch it.
1784 1784 */
1785 1785 idm_task_rele(itask->it_idm_task);
1786 1786 if (!aborted) {
1787 1787 stmf_send_status_done(itask->it_stmf_task,
1788 1788 iscsit_idm_to_stmf(pdu->isp_status), STMF_IOF_LPORT_DONE);
1789 1789 }
1790 1790 kmem_cache_free(iscsit_status_pdu_cache, pdu);
1791 1791 }
1792 1792
1793 1793 /*ARGSUSED*/
1794 1794 static void
1795 1795 iscsit_send_status_done(idm_pdu_t *pdu, idm_status_t status)
1796 1796 {
1797 1797 iscsit_task_t *itask;
1798 1798 boolean_t aborted;
1799 1799
1800 1800 itask = pdu->isp_private;
1801 1801 aborted = itask->it_stmf_abort;
1802 1802
1803 1803 /*
1804 1804 * After releasing the hold the task may be freed at any time so
1805 1805 * don't touch it.
1806 1806 */
1807 1807 idm_task_rele(itask->it_idm_task);
1808 1808 if (!aborted) {
1809 1809 stmf_send_status_done(itask->it_stmf_task,
1810 1810 iscsit_idm_to_stmf(pdu->isp_status), STMF_IOF_LPORT_DONE);
1811 1811 }
1812 1812 idm_pdu_free(pdu);
1813 1813 }
1814 1814
1815 1815
1816 1816 void
1817 1817 iscsit_lport_task_free(scsi_task_t *task)
1818 1818 {
1819 1819 iscsit_task_t *itask = task->task_port_private;
1820 1820
1821 1821 /* We only call idm_task_start for regular tasks, not task management */
1822 1822 if (task->task_mgmt_function == TM_NONE) {
1823 1823 idm_task_done(itask->it_idm_task);
1824 1824 iscsit_task_free(itask);
1825 1825 return;
1826 1826 } else {
1827 1827 iscsit_tm_task_free(itask);
1828 1828 }
1829 1829 }
1830 1830
1831 1831 /*ARGSUSED*/
1832 1832 stmf_status_t
1833 1833 iscsit_abort(stmf_local_port_t *lport, int abort_cmd, void *arg, uint32_t flags)
1834 1834 {
1835 1835 scsi_task_t *st = (scsi_task_t *)arg;
1836 1836 iscsit_task_t *iscsit_task;
1837 1837 idm_task_t *idt;
1838 1838
1839 1839 /*
1840 1840 * If this is a task management request then there's really not much to
1841 1841 * do.
1842 1842 */
1843 1843 if (st->task_mgmt_function != TM_NONE) {
1844 1844 return (STMF_ABORT_SUCCESS);
1845 1845 }
1846 1846
1847 1847 /*
1848 1848 * Regular task, start cleaning up
1849 1849 */
1850 1850 iscsit_task = st->task_port_private;
1851 1851 idt = iscsit_task->it_idm_task;
1852 1852 mutex_enter(&iscsit_task->it_mutex);
1853 1853 iscsit_task->it_stmf_abort = B_TRUE;
1854 1854 if (iscsit_task->it_aborted) {
1855 1855 mutex_exit(&iscsit_task->it_mutex);
1856 1856 /*
1857 1857 * Task is no longer active
1858 1858 */
1859 1859 iscsit_task_done(iscsit_task);
1860 1860
1861 1861 /*
1862 1862 * STMF specification is wrong... says to return
1863 1863 * STMF_ABORTED, the code actually looks for
1864 1864 * STMF_ABORT_SUCCESS.
1865 1865 */
1866 1866 return (STMF_ABORT_SUCCESS);
1867 1867 } else {
1868 1868 mutex_exit(&iscsit_task->it_mutex);
1869 1869 /*
1870 1870 * Call IDM to abort the task. Due to a variety of
1871 1871 * circumstances the task may already be in the process of
1872 1872 * aborting.
1873 1873 * We'll let IDM worry about rationalizing all that except
1874 1874 * for one particular instance. If the state of the task
1875 1875 * is TASK_COMPLETE, we need to indicate to the framework
1876 1876 * that we are in fact done. This typically happens with
1877 1877 * framework-initiated task management type requests
1878 1878 * (e.g. abort task).
1879 1879 */
1880 1880 if (idt->idt_state == TASK_COMPLETE) {
1881 1881 idm_refcnt_wait_ref(&idt->idt_refcnt);
1882 1882 return (STMF_ABORT_SUCCESS);
1883 1883 } else {
1884 1884 idm_task_abort(idt->idt_ic, idt, AT_TASK_MGMT_ABORT);
1885 1885 return (STMF_SUCCESS);
1886 1886 }
1887 1887 }
1888 1888
1889 1889 /*NOTREACHED*/
1890 1890 }
1891 1891
1892 1892 /*ARGSUSED*/
1893 1893 void
1894 1894 iscsit_ctl(stmf_local_port_t *lport, int cmd, void *arg)
1895 1895 {
1896 1896 iscsit_tgt_t *iscsit_tgt;
1897 1897
1898 1898 ASSERT((cmd == STMF_CMD_LPORT_ONLINE) ||
1899 1899 (cmd == STMF_ACK_LPORT_ONLINE_COMPLETE) ||
1900 1900 (cmd == STMF_CMD_LPORT_OFFLINE) ||
1901 1901 (cmd == STMF_ACK_LPORT_OFFLINE_COMPLETE));
1902 1902
1903 1903 iscsit_tgt = (iscsit_tgt_t *)lport->lport_port_private;
1904 1904
1905 1905 switch (cmd) {
1906 1906 case STMF_CMD_LPORT_ONLINE:
1907 1907 iscsit_tgt_sm_event(iscsit_tgt, TE_STMF_ONLINE_REQ);
1908 1908 break;
1909 1909 case STMF_CMD_LPORT_OFFLINE:
1910 1910 iscsit_tgt_sm_event(iscsit_tgt, TE_STMF_OFFLINE_REQ);
1911 1911 break;
1912 1912 case STMF_ACK_LPORT_ONLINE_COMPLETE:
1913 1913 iscsit_tgt_sm_event(iscsit_tgt, TE_STMF_ONLINE_COMPLETE_ACK);
1914 1914 break;
1915 1915 case STMF_ACK_LPORT_OFFLINE_COMPLETE:
1916 1916 iscsit_tgt_sm_event(iscsit_tgt, TE_STMF_OFFLINE_COMPLETE_ACK);
1917 1917 break;
1918 1918
1919 1919 default:
1920 1920 break;
1921 1921 }
1922 1922 }
1923 1923
1924 1924 static stmf_status_t
1925 1925 iscsit_idm_to_stmf(idm_status_t idmrc)
1926 1926 {
1927 1927 switch (idmrc) {
1928 1928 case IDM_STATUS_SUCCESS:
1929 1929 return (STMF_SUCCESS);
1930 1930 default:
1931 1931 return (STMF_FAILURE);
1932 1932 }
1933 1933 /*NOTREACHED*/
1934 1934 }
1935 1935
1936 1936 void
1937 1937 iscsit_op_scsi_cmd(idm_conn_t *ic, idm_pdu_t *rx_pdu)
1938 1938 {
1939 1939 iscsit_conn_t *ict = ic->ic_handle;
1940 1940
1941 1941 if (iscsit_check_cmdsn_and_queue(rx_pdu)) {
1942 1942 iscsit_post_scsi_cmd(ic, rx_pdu);
1943 1943 }
1944 1944 iscsit_process_pdu_in_queue(ict->ict_sess);
1945 1945 }
1946 1946
1947 1947 /*
1948 1948 * ISCSI protocol
1949 1949 */
1950 1950
1951 1951 void
1952 1952 iscsit_post_scsi_cmd(idm_conn_t *ic, idm_pdu_t *rx_pdu)
1953 1953 {
1954 1954 iscsit_conn_t *ict;
1955 1955 iscsit_task_t *itask;
1956 1956 scsi_task_t *task;
1957 1957 iscsit_buf_t *ibuf;
1958 1958 iscsi_scsi_cmd_hdr_t *iscsi_scsi =
1959 1959 (iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr;
1960 1960 iscsi_addl_hdr_t *ahs_hdr;
1961 1961 uint16_t addl_cdb_len = 0;
1962 1962
1963 1963 ict = ic->ic_handle;
1964 1964
1965 1965 itask = iscsit_task_alloc(ict);
1966 1966 if (itask == NULL) {
1967 1967 /* Finish processing request */
1968 1968 iscsit_set_cmdsn(ict, rx_pdu);
1969 1969
1970 1970 iscsit_send_direct_scsi_resp(ict, rx_pdu,
1971 1971 ISCSI_STATUS_CMD_COMPLETED, STATUS_BUSY);
1972 1972 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
1973 1973 return;
1974 1974 }
1975 1975
1976 1976 /*
1977 1977 * Note CmdSN and ITT in task. IDM will have already validated this
1978 1978 * request against the connection state so we don't need to check
1979 1979 * that (the connection may have changed state in the meantime but
1980 1980 * we will catch that when we try to send a response)
1981 1981 */
1982 1982 itask->it_cmdsn = ntohl(iscsi_scsi->cmdsn);
1983 1983 itask->it_itt = iscsi_scsi->itt;
1984 1984
1985 1985 /*
1986 1986 * Check for extended CDB AHS
1987 1987 */
1988 1988 if (iscsi_scsi->hlength > 0) {
1989 1989 ahs_hdr = (iscsi_addl_hdr_t *)iscsi_scsi;
1990 1990 addl_cdb_len = ((ahs_hdr->ahs_hlen_hi << 8) |
1991 1991 ahs_hdr->ahs_hlen_lo) - 1; /* Adjust for reserved byte */
1992 1992 if (((addl_cdb_len + 4) / sizeof (uint32_t)) >
1993 1993 iscsi_scsi->hlength) {
1994 1994 /* Mangled header info, drop it */
1995 1995 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
1996 1996 return;
1997 1997 }
1998 1998 }
1999 1999
2000 2000 ict = rx_pdu->isp_ic->ic_handle; /* IDM client private */
2001 2001
2002 2002 /*
2003 2003 * Add task to session list. This function will also check to
2004 2004 * ensure that the task does not already exist.
2005 2005 */
2006 2006 if (iscsit_task_start(itask) != IDM_STATUS_SUCCESS) {
2007 2007 /*
2008 2008 * Task exists, free all resources and reject. Don't
2009 2009 * update expcmdsn in this case because RFC 3720 says
2010 2010 * "The CmdSN of the rejected command PDU (if it is a
2011 2011 * non-immediate command) MUST NOT be considered received
2012 2012 * by the target (i.e., a command sequence gap must be
2013 2013 * assumed for the CmdSN), even though the CmdSN of the
2014 2014 * rejected command PDU may be reliably ascertained. Upon
2015 2015 * receiving the Reject, the initiator MUST plug the CmdSN
2016 2016 * gap in order to continue to use the session. The gap
2017 2017 * may be plugged either by transmitting a command PDU
2018 2018 * with the same CmdSN, or by aborting the task (see section
2019 2019 * 6.9 on how an abort may plug a CmdSN gap)." (Section 6.3)
2020 2020 */
2021 2021 iscsit_task_free(itask);
2022 2022 iscsit_send_reject(ict, rx_pdu, ISCSI_REJECT_TASK_IN_PROGRESS);
2023 2023 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2024 2024 return;
2025 2025 }
2026 2026
2027 2027 /* Update sequence numbers */
2028 2028 iscsit_set_cmdsn(ict, rx_pdu);
2029 2029
2030 2030 /*
2031 2031 * Allocate STMF task
2032 2032 */
2033 2033 itask->it_stmf_task = stmf_task_alloc(
2034 2034 itask->it_ict->ict_sess->ist_lport,
2035 2035 itask->it_ict->ict_sess->ist_stmf_sess, iscsi_scsi->lun,
2036 2036 16 + addl_cdb_len, 0);
2037 2037 if (itask->it_stmf_task == NULL) {
2038 2038 /*
2039 2039 * Either stmf really couldn't get memory for a task or,
2040 2040 * more likely, the LU is currently in reset. Either way
2041 2041 * we have no choice but to fail the request.
2042 2042 */
2043 2043 iscsit_task_done(itask);
2044 2044 iscsit_task_free(itask);
2045 2045 iscsit_send_direct_scsi_resp(ict, rx_pdu,
2046 2046 ISCSI_STATUS_CMD_COMPLETED, STATUS_BUSY);
2047 2047 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2048 2048 return;
2049 2049 }
2050 2050
2051 2051 task = itask->it_stmf_task;
2052 2052 task->task_port_private = itask;
2053 2053
2054 2054 bcopy(iscsi_scsi->lun, task->task_lun_no, sizeof (task->task_lun_no));
2055 2055
2056 2056 /*
2057 2057 * iSCSI and Comstar use the same values. Should we rely on this
2058 2058 * or translate them bit-wise?
2059 2059 */
2060 2060
2061 2061 task->task_flags =
2062 2062 (((iscsi_scsi->flags & ISCSI_FLAG_CMD_READ) ? TF_READ_DATA : 0) |
2063 2063 ((iscsi_scsi->flags & ISCSI_FLAG_CMD_WRITE) ? TF_WRITE_DATA : 0) |
2064 2064 ((rx_pdu->isp_datalen == 0) ? 0 : TF_INITIAL_BURST));
2065 2065
2066 2066 switch (iscsi_scsi->flags & ISCSI_FLAG_CMD_ATTR_MASK) {
2067 2067 case ISCSI_ATTR_UNTAGGED:
2068 2068 break;
2069 2069 case ISCSI_ATTR_SIMPLE:
2070 2070 task->task_additional_flags |= TF_ATTR_SIMPLE_QUEUE;
2071 2071 break;
2072 2072 case ISCSI_ATTR_ORDERED:
2073 2073 task->task_additional_flags |= TF_ATTR_ORDERED_QUEUE;
2074 2074 break;
2075 2075 case ISCSI_ATTR_HEAD_OF_QUEUE:
2076 2076 task->task_additional_flags |= TF_ATTR_HEAD_OF_QUEUE;
2077 2077 break;
2078 2078 case ISCSI_ATTR_ACA:
2079 2079 task->task_additional_flags |= TF_ATTR_ACA;
2080 2080 break;
2081 2081 default:
2082 2082 /* Protocol error but just take it, treat as untagged */
2083 2083 break;
2084 2084 }
2085 2085
2086 2086
2087 2087 task->task_additional_flags = 0;
2088 2088 task->task_priority = 0;
2089 2089 task->task_mgmt_function = TM_NONE;
2090 2090
2091 2091 /*
2092 2092 * This "task_max_nbufs" doesn't map well to BIDI. We probably need
2093 2093 * parameter for each direction. "MaxOutstandingR2T" may very well
2094 2094 * be set to one which could prevent us from doing simultaneous
2095 2095 * transfers in each direction.
2096 2096 */
2097 2097 task->task_max_nbufs = (iscsi_scsi->flags & ISCSI_FLAG_CMD_WRITE) ?
2098 2098 ict->ict_op.op_max_outstanding_r2t : STMF_BUFS_MAX;
2099 2099 task->task_cmd_seq_no = ntohl(iscsi_scsi->itt);
2100 2100 task->task_expected_xfer_length = ntohl(iscsi_scsi->data_length);
2101 2101
2102 2102 /* Copy CDB */
2103 2103 bcopy(iscsi_scsi->scb, task->task_cdb, 16);
2104 2104 if (addl_cdb_len > 0) {
2105 2105 bcopy(ahs_hdr->ahs_extscb, task->task_cdb + 16, addl_cdb_len);
2106 2106 }
2107 2107
2108 2108 DTRACE_ISCSI_3(scsi__command, idm_conn_t *, ic,
2109 2109 iscsi_scsi_cmd_hdr_t *, (iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr,
2110 2110 scsi_task_t *, task);
2111 2111
2112 2112 /*
2113 2113 * Copy the transport header into the task handle from the PDU
2114 2114 * handle. The transport header describes this task's remote tagged
2115 2115 * buffer.
2116 2116 */
2117 2117 if (rx_pdu->isp_transport_hdrlen != 0) {
2118 2118 bcopy(rx_pdu->isp_transport_hdr,
2119 2119 itask->it_idm_task->idt_transport_hdr,
2120 2120 rx_pdu->isp_transport_hdrlen);
2121 2121 }
2122 2122
2123 2123 /*
2124 2124 * Tell IDM about our new active task
2125 2125 */
2126 2126 idm_task_start(itask->it_idm_task, (uintptr_t)itask->it_itt);
2127 2127
2128 2128 /*
2129 2129 * If we have any immediate data then setup the immediate buffer
2130 2130 * context that comes with the task
2131 2131 */
2132 2132 if (rx_pdu->isp_datalen) {
2133 2133 ibuf = itask->it_immed_data;
2134 2134 ibuf->ibuf_immed_data_pdu = rx_pdu;
2135 2135 ibuf->ibuf_stmf_buf->db_data_size = rx_pdu->isp_datalen;
2136 2136 ibuf->ibuf_stmf_buf->db_buf_size = rx_pdu->isp_datalen;
2137 2137 ibuf->ibuf_stmf_buf->db_relative_offset = 0;
2138 2138 ibuf->ibuf_stmf_buf->db_sglist[0].seg_length =
2139 2139 rx_pdu->isp_datalen;
2140 2140 ibuf->ibuf_stmf_buf->db_sglist[0].seg_addr = rx_pdu->isp_data;
2141 2141
2142 2142 DTRACE_ISCSI_8(xfer__start, idm_conn_t *, ic,
2143 2143 uintptr_t, ibuf->ibuf_stmf_buf->db_sglist[0].seg_addr,
2144 2144 uint32_t, ibuf->ibuf_stmf_buf->db_relative_offset,
2145 2145 uint64_t, 0, uint32_t, 0, uint32_t, 0, /* no raddr */
2146 2146 uint32_t, rx_pdu->isp_datalen, int, XFER_BUF_TX_TO_INI);
2147 2147
2148 2148 /*
2149 2149 * For immediate data transfer, there is no callback from
2150 2150 * stmf to indicate that the initial burst of data is
2151 2151 * transferred successfully. In some cases, the task can
2152 2152 * get freed before execution returns from stmf_post_task.
2153 2153 * Although this xfer-start/done probe accurately tracks
2154 2154 * the size of the transfer, it does only provide a best
2155 2155 * effort on the timing of the transfer.
2156 2156 */
2157 2157 DTRACE_ISCSI_8(xfer__done, idm_conn_t *, ic,
2158 2158 uintptr_t, ibuf->ibuf_stmf_buf->db_sglist[0].seg_addr,
2159 2159 uint32_t, ibuf->ibuf_stmf_buf->db_relative_offset,
2160 2160 uint64_t, 0, uint32_t, 0, uint32_t, 0, /* no raddr */
2161 2161 uint32_t, rx_pdu->isp_datalen, int, XFER_BUF_TX_TO_INI);
2162 2162 stmf_post_task(task, ibuf->ibuf_stmf_buf);
2163 2163 } else {
2164 2164
2165 2165 stmf_post_task(task, NULL);
2166 2166 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2167 2167 }
2168 2168 }
2169 2169
2170 2170 void
2171 2171 iscsit_deferred_dispatch(idm_pdu_t *rx_pdu)
2172 2172 {
2173 2173 iscsit_conn_t *ict = rx_pdu->isp_ic->ic_handle;
2174 2174
2175 2175 /*
2176 2176 * If this isn't a login packet, we need a session. Otherwise
2177 2177 * this is a protocol error (perhaps one IDM should've caught?).
2178 2178 */
2179 2179 if (IDM_PDU_OPCODE(rx_pdu) != ISCSI_OP_LOGIN_CMD &&
2180 2180 ict->ict_sess == NULL) {
2181 2181 DTRACE_PROBE2(iscsi__idm__deferred__no__session,
2182 2182 iscsit_conn_t *, ict, idm_pdu_t *, rx_pdu);
2183 2183 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
2184 2184 return;
2185 2185 }
2186 2186
2187 2187 /*
2188 2188 * If the connection has been lost then ignore new PDU's
2189 2189 */
2190 2190 mutex_enter(&ict->ict_mutex);
2191 2191 if (ict->ict_lost) {
2192 2192 mutex_exit(&ict->ict_mutex);
2193 2193 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
2194 2194 return;
2195 2195 }
2196 2196
2197 2197 /*
2198 2198 * Grab a hold on the connection to prevent it from going away
2199 2199 * between now and when the taskq function is called.
2200 2200 */
2201 2201 iscsit_conn_dispatch_hold(ict);
2202 2202 mutex_exit(&ict->ict_mutex);
2203 2203
2204 2204 taskq_dispatch_ent(iscsit_global.global_dispatch_taskq,
2205 2205 iscsit_deferred, rx_pdu, 0, &rx_pdu->isp_tqent);
2206 2206 }
2207 2207
2208 2208 static void
2209 2209 iscsit_deferred(void *rx_pdu_void)
2210 2210 {
2211 2211 idm_pdu_t *rx_pdu = rx_pdu_void;
2212 2212 idm_conn_t *ic = rx_pdu->isp_ic;
2213 2213 iscsit_conn_t *ict = ic->ic_handle;
2214 2214
2215 2215 /*
2216 2216 * NOP and Task Management Commands can be marked for immediate
2217 2217 * delivery. Commands marked as 'Immediate' are to be considered
2218 2218 * for execution as soon as they arrive on the target. So these
2219 2219 * should not be checked for sequence order and put in a queue.
2220 2220 * The CmdSN is not advanced for Immediate Commands.
2221 2221 */
2222 2222 switch (IDM_PDU_OPCODE(rx_pdu)) {
2223 2223 case ISCSI_OP_NOOP_OUT:
2224 2224 if (iscsit_check_cmdsn_and_queue(rx_pdu)) {
2225 2225 iscsit_set_cmdsn(ict, rx_pdu);
2226 2226 iscsit_pdu_op_noop(ict, rx_pdu);
2227 2227 }
2228 2228 break;
2229 2229 case ISCSI_OP_LOGIN_CMD:
2230 2230 iscsit_pdu_op_login_cmd(ict, rx_pdu);
2231 2231 iscsit_conn_dispatch_rele(ict);
2232 2232 return;
2233 2233 case ISCSI_OP_TEXT_CMD:
2234 2234 if (iscsit_check_cmdsn_and_queue(rx_pdu)) {
2235 2235 iscsit_set_cmdsn(ict, rx_pdu);
2236 2236 iscsit_pdu_op_text_cmd(ict, rx_pdu);
2237 2237 }
2238 2238 break;
2239 2239 case ISCSI_OP_LOGOUT_CMD:
2240 2240 if (iscsit_check_cmdsn_and_queue(rx_pdu)) {
2241 2241 iscsit_set_cmdsn(ict, rx_pdu);
2242 2242 iscsit_pdu_op_logout_cmd(ict, rx_pdu);
2243 2243 }
2244 2244 break;
2245 2245 default:
2246 2246 /* Protocol error. IDM should have caught this */
2247 2247 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
2248 2248 ASSERT(0);
2249 2249 break;
2250 2250 }
2251 2251 /*
2252 2252 * Check if there are other PDUs in the session staging queue
2253 2253 * waiting to be posted to SCSI layer.
2254 2254 */
2255 2255 iscsit_process_pdu_in_queue(ict->ict_sess);
2256 2256
2257 2257 iscsit_conn_dispatch_rele(ict);
2258 2258 }
2259 2259
2260 2260 static void
2261 2261 iscsit_send_direct_scsi_resp(iscsit_conn_t *ict, idm_pdu_t *rx_pdu,
2262 2262 uint8_t response, uint8_t cmd_status)
2263 2263 {
2264 2264 idm_pdu_t *rsp_pdu;
2265 2265 idm_conn_t *ic;
2266 2266 iscsi_scsi_rsp_hdr_t *resp;
2267 2267 iscsi_scsi_cmd_hdr_t *req =
2268 2268 (iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr;
2269 2269
2270 2270 ic = ict->ict_ic;
2271 2271
2272 2272 rsp_pdu = idm_pdu_alloc(sizeof (iscsi_scsi_rsp_hdr_t), 0);
2273 2273 idm_pdu_init(rsp_pdu, ic, NULL, NULL);
2274 2274 /*
2275 2275 * StatSN is incremented by 1 for every response sent on
2276 2276 * a connection except for responses sent as a result of
2277 2277 * a retry or SNACK
2278 2278 */
2279 2279 rsp_pdu->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
2280 2280
2281 2281 resp = (iscsi_scsi_rsp_hdr_t *)rsp_pdu->isp_hdr;
2282 2282
2283 2283 resp->opcode = ISCSI_OP_SCSI_RSP;
2284 2284 resp->flags = ISCSI_FLAG_FINAL;
2285 2285 resp->response = response;
2286 2286 resp->cmd_status = cmd_status;
2287 2287 resp->itt = req->itt;
2288 2288 if ((response == ISCSI_STATUS_CMD_COMPLETED) &&
2289 2289 (req->data_length != 0) &&
2290 2290 ((req->flags & ISCSI_FLAG_CMD_READ) ||
2291 2291 (req->flags & ISCSI_FLAG_CMD_WRITE))) {
2292 2292 resp->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
2293 2293 resp->residual_count = req->data_length;
2294 2294 }
2295 2295
2296 2296 DTRACE_PROBE4(iscsi__scsi__direct__response,
2297 2297 iscsit_conn_t *, ict,
2298 2298 uint8_t, resp->response,
2299 2299 uint8_t, resp->cmd_status,
2300 2300 idm_pdu_t *, rsp_pdu);
2301 2301
2302 2302 iscsit_pdu_tx(rsp_pdu);
2303 2303 }
2304 2304
2305 2305 void
2306 2306 iscsit_send_task_mgmt_resp(idm_pdu_t *tm_resp_pdu, uint8_t tm_status)
2307 2307 {
2308 2308 iscsi_scsi_task_mgt_rsp_hdr_t *tm_resp;
2309 2309
2310 2310 /*
2311 2311 * The target must take note of the last-sent StatSN.
2312 2312 * The StatSN is to be incremented after sending a
2313 2313 * task management response. Digest recovery can only
2314 2314 * work if StatSN is incremented.
2315 2315 */
2316 2316 tm_resp_pdu->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
2317 2317 tm_resp = (iscsi_scsi_task_mgt_rsp_hdr_t *)tm_resp_pdu->isp_hdr;
2318 2318 tm_resp->response = tm_status;
2319 2319
2320 2320 DTRACE_PROBE3(iscsi__scsi__tm__response,
2321 2321 iscsit_conn_t *, tm_resp_pdu->isp_ic->ic_handle,
2322 2322 uint8_t, tm_resp->response,
2323 2323 idm_pdu_t *, tm_resp_pdu);
2324 2324 iscsit_pdu_tx(tm_resp_pdu);
2325 2325 }
2326 2326
2327 2327 void
2328 2328 iscsit_op_scsi_task_mgmt(iscsit_conn_t *ict, idm_pdu_t *rx_pdu)
2329 2329 {
2330 2330 idm_pdu_t *tm_resp_pdu;
2331 2331 iscsit_task_t *itask;
2332 2332 iscsit_task_t *tm_itask;
2333 2333 scsi_task_t *task;
2334 2334 iscsi_scsi_task_mgt_hdr_t *iscsi_tm =
2335 2335 (iscsi_scsi_task_mgt_hdr_t *)rx_pdu->isp_hdr;
2336 2336 iscsi_scsi_task_mgt_rsp_hdr_t *iscsi_tm_rsp =
2337 2337 (iscsi_scsi_task_mgt_rsp_hdr_t *)rx_pdu->isp_hdr;
2338 2338 uint32_t rtt, cmdsn, refcmdsn;
2339 2339 uint8_t tm_func;
2340 2340
2341 2341 /*
2342 2342 * Setup response PDU (response field will get filled in later)
2343 2343 */
2344 2344 tm_resp_pdu = idm_pdu_alloc(sizeof (iscsi_scsi_task_mgt_rsp_hdr_t), 0);
2345 2345 if (tm_resp_pdu == NULL) {
2346 2346 /* Can't respond, just drop it */
2347 2347 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2348 2348 return;
2349 2349 }
2350 2350 idm_pdu_init(tm_resp_pdu, ict->ict_ic, NULL, NULL);
2351 2351 iscsi_tm_rsp = (iscsi_scsi_task_mgt_rsp_hdr_t *)tm_resp_pdu->isp_hdr;
2352 2352 bzero(iscsi_tm_rsp, sizeof (iscsi_scsi_task_mgt_rsp_hdr_t));
2353 2353 iscsi_tm_rsp->opcode = ISCSI_OP_SCSI_TASK_MGT_RSP;
2354 2354 iscsi_tm_rsp->flags = ISCSI_FLAG_FINAL;
2355 2355 iscsi_tm_rsp->itt = rx_pdu->isp_hdr->itt;
2356 2356
2357 2357 /*
2358 2358 * Figure out what we're being asked to do.
2359 2359 */
2360 2360 DTRACE_PROBE4(iscsi__scsi__tm__request,
2361 2361 iscsit_conn_t *, ict,
2362 2362 uint8_t, (iscsi_tm->function & ISCSI_FLAG_TASK_MGMT_FUNCTION_MASK),
2363 2363 uint32_t, iscsi_tm->rtt,
2364 2364 idm_pdu_t *, rx_pdu);
2365 2365 switch (iscsi_tm->function & ISCSI_FLAG_TASK_MGMT_FUNCTION_MASK) {
2366 2366 case ISCSI_TM_FUNC_ABORT_TASK:
2367 2367 /*
2368 2368 * STMF doesn't currently support the "abort task" task
2369 2369 * management command although it does support aborting
2370 2370 * an individual task. We'll get STMF to abort the task
2371 2371 * for us but handle the details of the task management
2372 2372 * command ourselves.
2373 2373 *
2374 2374 * Find the task associated with the referenced task tag.
2375 2375 */
2376 2376 rtt = iscsi_tm->rtt;
2377 2377 itask = (iscsit_task_t *)idm_task_find_by_handle(ict->ict_ic,
2378 2378 (uintptr_t)rtt);
2379 2379
2380 2380 if (itask == NULL) {
2381 2381 cmdsn = ntohl(iscsi_tm->cmdsn);
2382 2382 refcmdsn = ntohl(iscsi_tm->refcmdsn);
2383 2383
2384 2384 /*
2385 2385 * Task was not found. But the SCSI command could be
2386 2386 * on the rxpdu wait queue. If RefCmdSN is within
2387 2387 * the CmdSN window and less than CmdSN of the TM
2388 2388 * function, return "Function Complete". Otherwise,
2389 2389 * return "Task Does Not Exist".
2390 2390 */
2391 2391
2392 2392 if (iscsit_cmdsn_in_window(ict, refcmdsn) &&
2393 2393 iscsit_sna_lt(refcmdsn, cmdsn)) {
2394 2394 mutex_enter(&ict->ict_sess->ist_sn_mutex);
2395 2395 (void) iscsit_remove_pdu_from_queue(
2396 2396 ict->ict_sess, refcmdsn);
2397 2397 iscsit_conn_dispatch_rele(ict);
2398 2398 mutex_exit(&ict->ict_sess->ist_sn_mutex);
2399 2399 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2400 2400 SCSI_TCP_TM_RESP_COMPLETE);
2401 2401 } else {
2402 2402 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2403 2403 SCSI_TCP_TM_RESP_NO_TASK);
2404 2404 }
2405 2405 } else {
2406 2406
2407 2407 /*
2408 2408 * Tell STMF to abort the task. This will do no harm
2409 2409 * if the task is already complete.
2410 2410 */
2411 2411 stmf_abort(STMF_QUEUE_TASK_ABORT, itask->it_stmf_task,
2412 2412 STMF_ABORTED, NULL);
2413 2413
2414 2414 /*
2415 2415 * Make sure the task hasn't already completed
2416 2416 */
2417 2417 mutex_enter(&itask->it_idm_task->idt_mutex);
2418 2418 if ((itask->it_idm_task->idt_state == TASK_COMPLETE) ||
2419 2419 (itask->it_idm_task->idt_state == TASK_IDLE)) {
2420 2420 /*
2421 2421 * Task is complete, return "Task Does Not
2422 2422 * Exist"
2423 2423 */
2424 2424 mutex_exit(&itask->it_idm_task->idt_mutex);
2425 2425 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2426 2426 SCSI_TCP_TM_RESP_NO_TASK);
2427 2427 } else {
2428 2428 /*
2429 2429 * STMF is now aborting the task, return
2430 2430 * "Function Complete"
2431 2431 */
2432 2432 mutex_exit(&itask->it_idm_task->idt_mutex);
2433 2433 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2434 2434 SCSI_TCP_TM_RESP_COMPLETE);
2435 2435 }
2436 2436 idm_task_rele(itask->it_idm_task);
2437 2437 }
2438 2438 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2439 2439 return;
2440 2440
2441 2441 case ISCSI_TM_FUNC_ABORT_TASK_SET:
2442 2442 tm_func = TM_ABORT_TASK_SET;
2443 2443 break;
2444 2444
2445 2445 case ISCSI_TM_FUNC_CLEAR_ACA:
2446 2446 tm_func = TM_CLEAR_ACA;
2447 2447 break;
2448 2448
2449 2449 case ISCSI_TM_FUNC_CLEAR_TASK_SET:
2450 2450 tm_func = TM_CLEAR_TASK_SET;
2451 2451 break;
2452 2452
2453 2453 case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET:
2454 2454 tm_func = TM_LUN_RESET;
2455 2455 break;
2456 2456
2457 2457 case ISCSI_TM_FUNC_TARGET_WARM_RESET:
2458 2458 tm_func = TM_TARGET_WARM_RESET;
2459 2459 break;
2460 2460
2461 2461 case ISCSI_TM_FUNC_TARGET_COLD_RESET:
2462 2462 tm_func = TM_TARGET_COLD_RESET;
2463 2463 break;
2464 2464
2465 2465 case ISCSI_TM_FUNC_TASK_REASSIGN:
2466 2466 /*
2467 2467 * We do not currently support allegiance reassignment. When
2468 2468 * we start supporting ERL1+, we will need to.
2469 2469 */
2470 2470 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2471 2471 SCSI_TCP_TM_RESP_NO_ALLG_REASSN);
2472 2472 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2473 2473 return;
2474 2474
2475 2475 default:
2476 2476 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2477 2477 SCSI_TCP_TM_RESP_REJECTED);
2478 2478 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2479 2479 return;
2480 2480 }
2481 2481
2482 2482 tm_itask = iscsit_tm_task_alloc(ict);
2483 2483 if (tm_itask == NULL) {
2484 2484 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2485 2485 SCSI_TCP_TM_RESP_REJECTED);
2486 2486 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2487 2487 return;
2488 2488 }
2489 2489
2490 2490
2491 2491 task = stmf_task_alloc(ict->ict_sess->ist_lport,
2492 2492 ict->ict_sess->ist_stmf_sess, iscsi_tm->lun,
2493 2493 0, STMF_TASK_EXT_NONE);
2494 2494 if (task == NULL) {
2495 2495 /*
2496 2496 * If this happens, either the LU is in reset, couldn't
2497 2497 * get memory, or some other condition in which we simply
2498 2498 * can't complete this request. It would be nice to return
2499 2499 * an error code like "busy" but the closest we have is
2500 2500 * "rejected".
2501 2501 */
2502 2502 iscsit_send_task_mgmt_resp(tm_resp_pdu,
2503 2503 SCSI_TCP_TM_RESP_REJECTED);
2504 2504 iscsit_tm_task_free(tm_itask);
2505 2505 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2506 2506 return;
2507 2507 }
2508 2508
2509 2509 tm_itask->it_tm_pdu = tm_resp_pdu;
2510 2510 tm_itask->it_stmf_task = task;
2511 2511 task->task_port_private = tm_itask;
2512 2512 task->task_mgmt_function = tm_func;
2513 2513 task->task_additional_flags = TASK_AF_NO_EXPECTED_XFER_LENGTH;
2514 2514 task->task_priority = 0;
2515 2515 task->task_max_nbufs = STMF_BUFS_MAX;
2516 2516 task->task_cmd_seq_no = iscsi_tm->itt;
2517 2517 task->task_expected_xfer_length = 0;
2518 2518
2519 2519 stmf_post_task(task, NULL);
2520 2520 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2521 2521 }
2522 2522
2523 2523 static void
2524 2524 iscsit_pdu_op_noop(iscsit_conn_t *ict, idm_pdu_t *rx_pdu)
2525 2525 {
2526 2526 iscsi_nop_out_hdr_t *out = (iscsi_nop_out_hdr_t *)rx_pdu->isp_hdr;
2527 2527 iscsi_nop_in_hdr_t *in;
2528 2528 int resp_datalen;
2529 2529 idm_pdu_t *resp;
2530 2530
2531 2531 /* Ignore the response from initiator */
2532 2532 if ((out->itt == ISCSI_RSVD_TASK_TAG) ||
2533 2533 (out->ttt != ISCSI_RSVD_TASK_TAG)) {
2534 2534 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2535 2535 return;
2536 2536 }
2537 2537
2538 2538 /* Allocate a PDU to respond */
2539 2539 resp_datalen = ntoh24(out->dlength);
2540 2540 resp = idm_pdu_alloc(sizeof (iscsi_hdr_t), resp_datalen);
2541 2541 idm_pdu_init(resp, ict->ict_ic, NULL, NULL);
2542 2542 if (resp_datalen > 0) {
2543 2543 bcopy(rx_pdu->isp_data, resp->isp_data, resp_datalen);
2544 2544 }
2545 2545
2546 2546 /*
2547 2547 * When sending a NOP-In as a response to a NOP-Out from the initiator,
2548 2548 * the target must respond with the same initiator task tag that was
2549 2549 * provided in the NOP-Out request, the target transfer tag must be
2550 2550 * ISCSI_RSVD_TASK_TAG (0xffffffff) and StatSN will contain the next
2551 2551 * status sequence number. The StatSN for the connection is advanced
2552 2552 * after this PDU is sent.
2553 2553 */
2554 2554 in = (iscsi_nop_in_hdr_t *)resp->isp_hdr;
2555 2555 bzero(in, sizeof (*in));
2556 2556 in->opcode = ISCSI_OP_NOOP_IN;
2557 2557 in->flags = ISCSI_FLAG_FINAL;
2558 2558 bcopy(out->lun, in->lun, 8);
2559 2559 in->itt = out->itt;
2560 2560 in->ttt = ISCSI_RSVD_TASK_TAG;
2561 2561 hton24(in->dlength, resp_datalen);
2562 2562 resp->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
2563 2563 /* Any other field in resp to be set? */
2564 2564 iscsit_pdu_tx(resp);
2565 2565 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2566 2566 }
2567 2567
2568 2568 static void
2569 2569 iscsit_pdu_op_login_cmd(iscsit_conn_t *ict, idm_pdu_t *rx_pdu)
2570 2570 {
2571 2571
2572 2572 /*
2573 2573 * Submit PDU to login state machine. State machine will free the
2574 2574 * PDU.
2575 2575 */
2576 2576 iscsit_login_sm_event(ict, ILE_LOGIN_RCV, rx_pdu);
2577 2577 }
2578 2578
2579 2579 void
2580 2580 iscsit_pdu_op_logout_cmd(iscsit_conn_t *ict, idm_pdu_t *rx_pdu)
2581 2581 {
2582 2582 iscsi_logout_hdr_t *logout_req =
2583 2583 (iscsi_logout_hdr_t *)rx_pdu->isp_hdr;
2584 2584 iscsi_logout_rsp_hdr_t *logout_rsp;
2585 2585 idm_pdu_t *resp;
2586 2586
2587 2587 /* Allocate a PDU to respond */
2588 2588 resp = idm_pdu_alloc(sizeof (iscsi_hdr_t), 0);
2589 2589 idm_pdu_init(resp, ict->ict_ic, NULL, NULL);
2590 2590 /*
2591 2591 * The StatSN is to be sent to the initiator,
2592 2592 * it is not required to increment the number
2593 2593 * as the connection is terminating.
2594 2594 */
2595 2595 resp->isp_flags |= IDM_PDU_SET_STATSN;
2596 2596 /*
2597 2597 * Logout results in the immediate termination of all tasks except
2598 2598 * if the logout reason is ISCSI_LOGOUT_REASON_RECOVERY. The
2599 2599 * connection state machine will drive this task cleanup automatically
2600 2600 * so we don't need to handle that here.
2601 2601 */
2602 2602 logout_rsp = (iscsi_logout_rsp_hdr_t *)resp->isp_hdr;
2603 2603 bzero(logout_rsp, sizeof (*logout_rsp));
2604 2604 logout_rsp->opcode = ISCSI_OP_LOGOUT_RSP;
2605 2605 logout_rsp->flags = ISCSI_FLAG_FINAL;
2606 2606 logout_rsp->itt = logout_req->itt;
2607 2607 if ((logout_req->flags & ISCSI_FLAG_LOGOUT_REASON_MASK) >
2608 2608 ISCSI_LOGOUT_REASON_RECOVERY) {
2609 2609 logout_rsp->response = ISCSI_LOGOUT_RECOVERY_UNSUPPORTED;
2610 2610 } else {
2611 2611 logout_rsp->response = ISCSI_LOGOUT_SUCCESS;
2612 2612 }
2613 2613
2614 2614 iscsit_pdu_tx(resp);
2615 2615 idm_pdu_complete(rx_pdu, IDM_STATUS_SUCCESS);
2616 2616 }
2617 2617
2618 2618 /*
2619 2619 * Calculate the number of outstanding commands we can process
2620 2620 */
2621 2621 int
2622 2622 iscsit_cmd_window()
2623 2623 {
2624 2624 /*
2625 2625 * Instead of using a pre-defined constant for the command window,
2626 2626 * it should be made confiurable and dynamic. With MC/S, sequence
2627 2627 * numbers will be used up at a much faster rate than with SC/S.
2628 2628 */
2629 2629 return (ISCSIT_MAX_WINDOW);
2630 2630 }
2631 2631
2632 2632 /*
2633 2633 * Set local registers based on incoming PDU
2634 2634 */
2635 2635 void
2636 2636 iscsit_set_cmdsn(iscsit_conn_t *ict, idm_pdu_t *rx_pdu)
2637 2637 {
2638 2638 iscsit_sess_t *ist;
2639 2639 iscsi_scsi_cmd_hdr_t *req;
2640 2640
2641 2641 ist = ict->ict_sess;
2642 2642
2643 2643 req = (iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr;
2644 2644 if (req->opcode & ISCSI_OP_IMMEDIATE) {
2645 2645 /* no cmdsn increment for immediate PDUs */
2646 2646 return;
2647 2647 }
2648 2648
2649 2649 /* Ensure that the ExpCmdSN advances in an orderly manner */
2650 2650 mutex_enter(&ist->ist_sn_mutex);
2651 2651 ist->ist_expcmdsn = ntohl(req->cmdsn) + 1;
2652 2652 ist->ist_maxcmdsn = ntohl(req->cmdsn) + iscsit_cmd_window();
2653 2653 mutex_exit(&ist->ist_sn_mutex);
2654 2654 }
2655 2655
2656 2656 /*
2657 2657 * Wrapper funtion, calls iscsi_calc_rspsn and idm_pdu_tx
2658 2658 */
2659 2659 void
2660 2660 iscsit_pdu_tx(idm_pdu_t *pdu)
2661 2661 {
2662 2662 iscsit_conn_t *ict = pdu->isp_ic->ic_handle;
2663 2663 iscsi_scsi_rsp_hdr_t *rsp = (iscsi_scsi_rsp_hdr_t *)pdu->isp_hdr;
2664 2664 iscsit_sess_t *ist = ict->ict_sess;
2665 2665
2666 2666 /*
2667 2667 * The command sequence numbers are session-wide and must stay
2668 2668 * consistent across the transfer, so protect the cmdsn with a
2669 2669 * mutex lock on the session. The status sequence number will
2670 2670 * be updated just before the transport layer transmits the PDU.
2671 2671 */
2672 2672
2673 2673 mutex_enter(&ict->ict_sess->ist_sn_mutex);
2674 2674 /* Set ExpCmdSN and MaxCmdSN */
2675 2675 rsp->maxcmdsn = htonl(ist->ist_maxcmdsn);
2676 2676 rsp->expcmdsn = htonl(ist->ist_expcmdsn);
2677 2677 idm_pdu_tx(pdu);
2678 2678 mutex_exit(&ict->ict_sess->ist_sn_mutex);
2679 2679 }
2680 2680
2681 2681 /*
2682 2682 * Internal functions
2683 2683 */
2684 2684
2685 2685 void
2686 2686 iscsit_send_async_event(iscsit_conn_t *ict, uint8_t event)
2687 2687 {
2688 2688 idm_pdu_t *abt;
2689 2689 iscsi_async_evt_hdr_t *async_abt;
2690 2690
2691 2691 /*
2692 2692 * Get a PDU to build the abort request.
2693 2693 */
2694 2694 abt = idm_pdu_alloc(sizeof (iscsi_hdr_t), 0);
2695 2695 if (abt == NULL) {
2696 2696 idm_conn_event(ict->ict_ic, CE_TRANSPORT_FAIL, NULL);
2697 2697 return;
2698 2698 }
2699 2699
2700 2700 /*
2701 2701 * A asynchronous message is sent by the target to request a logout.
2702 2702 * The StatSN for the connection is advanced after the PDU is sent
2703 2703 * to allow for initiator and target state synchronization.
2704 2704 */
2705 2705 idm_pdu_init(abt, ict->ict_ic, NULL, NULL);
2706 2706 abt->isp_datalen = 0;
2707 2707 abt->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
2708 2708
2709 2709 async_abt = (iscsi_async_evt_hdr_t *)abt->isp_hdr;
2710 2710 bzero(async_abt, sizeof (*async_abt));
2711 2711 async_abt->opcode = ISCSI_OP_ASYNC_EVENT;
2712 2712 async_abt->async_event = event;
2713 2713 async_abt->flags = ISCSI_FLAG_FINAL;
2714 2714 async_abt->rsvd4[0] = 0xff;
2715 2715 async_abt->rsvd4[1] = 0xff;
2716 2716 async_abt->rsvd4[2] = 0xff;
2717 2717 async_abt->rsvd4[3] = 0xff;
2718 2718
2719 2719 switch (event) {
2720 2720 case ISCSI_ASYNC_EVENT_REQUEST_LOGOUT:
2721 2721 async_abt->param3 = htons(IDM_LOGOUT_SECONDS);
2722 2722 break;
2723 2723 case ISCSI_ASYNC_EVENT_SCSI_EVENT:
2724 2724 case ISCSI_ASYNC_EVENT_DROPPING_CONNECTION:
2725 2725 case ISCSI_ASYNC_EVENT_DROPPING_ALL_CONNECTIONS:
2726 2726 case ISCSI_ASYNC_EVENT_PARAM_NEGOTIATION:
2727 2727 default:
2728 2728 ASSERT(0);
2729 2729 }
2730 2730
2731 2731 iscsit_pdu_tx(abt);
2732 2732 }
2733 2733
2734 2734 void
2735 2735 iscsit_send_reject(iscsit_conn_t *ict, idm_pdu_t *rejected_pdu, uint8_t reason)
2736 2736 {
2737 2737 idm_pdu_t *reject_pdu;
2738 2738 iscsi_reject_rsp_hdr_t *reject;
2739 2739
2740 2740 /*
2741 2741 * Get a PDU to build the abort request.
2742 2742 */
2743 2743 reject_pdu = idm_pdu_alloc(sizeof (iscsi_hdr_t),
2744 2744 rejected_pdu->isp_hdrlen);
2745 2745 if (reject_pdu == NULL) {
2746 2746 idm_conn_event(ict->ict_ic, CE_TRANSPORT_FAIL, NULL);
2747 2747 return;
2748 2748 }
2749 2749 idm_pdu_init(reject_pdu, ict->ict_ic, NULL, NULL);
2750 2750 /* StatSN is advanced after a Reject PDU */
2751 2751 reject_pdu->isp_flags |= IDM_PDU_SET_STATSN | IDM_PDU_ADVANCE_STATSN;
2752 2752 reject_pdu->isp_datalen = rejected_pdu->isp_hdrlen;
2753 2753 bcopy(rejected_pdu->isp_hdr, reject_pdu->isp_data,
2754 2754 rejected_pdu->isp_hdrlen);
2755 2755
2756 2756 reject = (iscsi_reject_rsp_hdr_t *)reject_pdu->isp_hdr;
2757 2757 bzero(reject, sizeof (*reject));
2758 2758 reject->opcode = ISCSI_OP_REJECT_MSG;
2759 2759 reject->reason = reason;
2760 2760 reject->flags = ISCSI_FLAG_FINAL;
2761 2761 hton24(reject->dlength, rejected_pdu->isp_hdrlen);
2762 2762 reject->must_be_ff[0] = 0xff;
2763 2763 reject->must_be_ff[1] = 0xff;
2764 2764 reject->must_be_ff[2] = 0xff;
2765 2765 reject->must_be_ff[3] = 0xff;
2766 2766
2767 2767 iscsit_pdu_tx(reject_pdu);
2768 2768 }
2769 2769
2770 2770
2771 2771 static iscsit_task_t *
2772 2772 iscsit_task_alloc(iscsit_conn_t *ict)
2773 2773 {
2774 2774 iscsit_task_t *itask;
2775 2775 iscsit_buf_t *immed_ibuf;
2776 2776
2777 2777 /*
2778 2778 * Possible items to pre-alloc if we cache iscsit_task_t's:
2779 2779 *
2780 2780 * Status PDU w/ sense buffer
2781 2781 * stmf_data_buf_t for immediate data
2782 2782 */
2783 2783 itask = kmem_alloc(sizeof (iscsit_task_t) + sizeof (iscsit_buf_t) +
2784 2784 sizeof (stmf_data_buf_t), KM_NOSLEEP);
2785 2785 if (itask != NULL) {
2786 2786 mutex_init(&itask->it_mutex, NULL, MUTEX_DRIVER, NULL);
2787 2787 itask->it_aborted = itask->it_stmf_abort =
2788 2788 itask->it_tm_task = 0;
2789 2789
2790 2790 immed_ibuf = (iscsit_buf_t *)(itask + 1);
2791 2791 bzero(immed_ibuf, sizeof (*immed_ibuf));
2792 2792 immed_ibuf->ibuf_is_immed = B_TRUE;
2793 2793 immed_ibuf->ibuf_stmf_buf = (stmf_data_buf_t *)(immed_ibuf + 1);
2794 2794
2795 2795 bzero(immed_ibuf->ibuf_stmf_buf, sizeof (stmf_data_buf_t));
2796 2796 immed_ibuf->ibuf_stmf_buf->db_port_private = immed_ibuf;
2797 2797 immed_ibuf->ibuf_stmf_buf->db_sglist_length = 1;
2798 2798 immed_ibuf->ibuf_stmf_buf->db_flags = DB_DIRECTION_FROM_RPORT |
2799 2799 DB_DONT_CACHE;
2800 2800 itask->it_immed_data = immed_ibuf;
2801 2801 itask->it_idm_task = idm_task_alloc(ict->ict_ic);
2802 2802 if (itask->it_idm_task != NULL) {
2803 2803 itask->it_idm_task->idt_private = itask;
2804 2804 itask->it_ict = ict;
2805 2805 itask->it_ttt = itask->it_idm_task->idt_tt;
2806 2806 return (itask);
2807 2807 } else {
2808 2808 kmem_free(itask, sizeof (iscsit_task_t) +
2809 2809 sizeof (iscsit_buf_t) + sizeof (stmf_data_buf_t));
2810 2810 }
2811 2811 }
2812 2812
2813 2813 return (NULL);
2814 2814 }
2815 2815
2816 2816 static void
2817 2817 iscsit_task_free(iscsit_task_t *itask)
2818 2818 {
2819 2819 idm_task_free(itask->it_idm_task);
2820 2820 mutex_destroy(&itask->it_mutex);
2821 2821 kmem_free(itask, sizeof (iscsit_task_t) +
2822 2822 sizeof (iscsit_buf_t) + sizeof (stmf_data_buf_t));
2823 2823 }
2824 2824
2825 2825 static iscsit_task_t *
2826 2826 iscsit_tm_task_alloc(iscsit_conn_t *ict)
2827 2827 {
2828 2828 iscsit_task_t *itask;
2829 2829
2830 2830 itask = kmem_zalloc(sizeof (iscsit_task_t), KM_NOSLEEP);
2831 2831 if (itask != NULL) {
2832 2832 idm_conn_hold(ict->ict_ic);
2833 2833 mutex_init(&itask->it_mutex, NULL, MUTEX_DRIVER, NULL);
2834 2834 itask->it_aborted = itask->it_stmf_abort =
2835 2835 itask->it_tm_responded = 0;
2836 2836 itask->it_tm_pdu = NULL;
2837 2837 itask->it_tm_task = 1;
2838 2838 itask->it_ict = ict;
2839 2839 }
2840 2840
2841 2841 return (itask);
2842 2842 }
2843 2843
2844 2844 static void
2845 2845 iscsit_tm_task_free(iscsit_task_t *itask)
2846 2846 {
2847 2847 /*
2848 2848 * If we responded then the call to idm_pdu_complete will free the
2849 2849 * PDU. Otherwise we got aborted before the TM function could
2850 2850 * complete and we need to free the PDU explicitly.
2851 2851 */
2852 2852 if (itask->it_tm_pdu != NULL && !itask->it_tm_responded)
2853 2853 idm_pdu_free(itask->it_tm_pdu);
2854 2854 idm_conn_rele(itask->it_ict->ict_ic);
2855 2855 mutex_destroy(&itask->it_mutex);
2856 2856 kmem_free(itask, sizeof (iscsit_task_t));
2857 2857 }
2858 2858
2859 2859 static idm_status_t
2860 2860 iscsit_task_start(iscsit_task_t *itask)
2861 2861 {
2862 2862 iscsit_sess_t *ist = itask->it_ict->ict_sess;
2863 2863 avl_index_t where;
2864 2864
2865 2865 /*
2866 2866 * Sanity check the ITT and ensure that this task does not already
2867 2867 * exist. If not then add the task to the session task list.
2868 2868 */
2869 2869 mutex_enter(&ist->ist_mutex);
2870 2870 mutex_enter(&itask->it_mutex);
2871 2871 itask->it_active = 1;
2872 2872 if (avl_find(&ist->ist_task_list, itask, &where) == NULL) {
2873 2873 /* New task, add to AVL */
2874 2874 avl_insert(&ist->ist_task_list, itask, where);
2875 2875 mutex_exit(&itask->it_mutex);
2876 2876 mutex_exit(&ist->ist_mutex);
2877 2877 return (IDM_STATUS_SUCCESS);
2878 2878 }
2879 2879 mutex_exit(&itask->it_mutex);
2880 2880 mutex_exit(&ist->ist_mutex);
2881 2881
2882 2882 return (IDM_STATUS_REJECT);
2883 2883 }
2884 2884
2885 2885 static void
2886 2886 iscsit_task_done(iscsit_task_t *itask)
2887 2887 {
2888 2888 iscsit_sess_t *ist = itask->it_ict->ict_sess;
2889 2889
2890 2890 mutex_enter(&ist->ist_mutex);
2891 2891 mutex_enter(&itask->it_mutex);
2892 2892 if (itask->it_active) {
2893 2893 avl_remove(&ist->ist_task_list, itask);
2894 2894 itask->it_active = 0;
2895 2895 }
2896 2896 mutex_exit(&itask->it_mutex);
2897 2897 mutex_exit(&ist->ist_mutex);
2898 2898 }
2899 2899
2900 2900 /*
2901 2901 * iscsit status PDU cache
2902 2902 */
2903 2903
2904 2904 /*ARGSUSED*/
2905 2905 static int
2906 2906 iscsit_status_pdu_constructor(void *pdu_void, void *arg, int flags)
2907 2907 {
2908 2908 idm_pdu_t *pdu = pdu_void;
2909 2909 iscsi_scsi_rsp_hdr_t *rsp;
2910 2910
2911 2911 bzero(pdu, sizeof (idm_pdu_t));
2912 2912 pdu->isp_callback = iscsit_send_good_status_done;
2913 2913 pdu->isp_magic = IDM_PDU_MAGIC;
2914 2914 pdu->isp_hdr = (iscsi_hdr_t *)(pdu + 1); /* Ptr arithmetic */
2915 2915 pdu->isp_hdrlen = sizeof (iscsi_hdr_t);
2916 2916
2917 2917 /* Setup status response */
2918 2918 rsp = (iscsi_scsi_rsp_hdr_t *)pdu->isp_hdr;
2919 2919 bzero(rsp, sizeof (*rsp));
2920 2920 rsp->opcode = ISCSI_OP_SCSI_RSP;
2921 2921 rsp->flags = ISCSI_FLAG_FINAL;
2922 2922 rsp->response = ISCSI_STATUS_CMD_COMPLETED;
2923 2923
2924 2924 return (0);
2925 2925 }
2926 2926
2927 2927 /*
2928 2928 * iscsit private data handler
2929 2929 */
2930 2930
2931 2931 /*ARGSUSED*/
2932 2932 static void
2933 2933 iscsit_pp_cb(struct stmf_port_provider *pp, int cmd, void *arg, uint32_t flags)
2934 2934 {
2935 2935 it_config_t *cfg;
2936 2936 nvlist_t *nvl;
2937 2937 iscsit_service_enabled_t old_state;
2938 2938
2939 2939 if ((cmd != STMF_PROVIDER_DATA_UPDATED) || (arg == NULL)) {
2940 2940 return;
2941 2941 }
2942 2942
2943 2943 nvl = (nvlist_t *)arg;
2944 2944
2945 2945 /* Translate nvlist */
2946 2946 if (it_nv_to_config(nvl, &cfg) != 0) {
2947 2947 cmn_err(CE_WARN, "Configuration is invalid");
2948 2948 return;
2949 2949 }
2950 2950
2951 2951 /* Check that no iSCSI ioctl is currently running */
2952 2952 mutex_enter(&iscsit_global.global_state_mutex);
2953 2953 old_state = iscsit_global.global_svc_state;
2954 2954 switch (iscsit_global.global_svc_state) {
2955 2955 case ISE_ENABLED:
2956 2956 case ISE_DISABLED:
2957 2957 iscsit_global.global_svc_state = ISE_BUSY;
2958 2958 break;
2959 2959 case ISE_ENABLING:
2960 2960 /*
2961 2961 * It is OK for the iscsit_pp_cb to be called from inside of
2962 2962 * an iSCSI ioctl only if we are currently executing inside
2963 2963 * of stmf_register_port_provider.
2964 2964 */
2965 2965 ASSERT((flags & STMF_PCB_PREG_COMPLETE) != 0);
2966 2966 break;
2967 2967 default:
2968 2968 cmn_err(CE_WARN, "iscsit_pp_cb called when global_svc_state"
2969 2969 " is not ENABLED(0x%x) -- ignoring",
2970 2970 iscsit_global.global_svc_state);
2971 2971 mutex_exit(&iscsit_global.global_state_mutex);
2972 2972 it_config_free_cmn(cfg);
2973 2973 return;
2974 2974 }
2975 2975 mutex_exit(&iscsit_global.global_state_mutex);
2976 2976
2977 2977 /* Update config */
2978 2978 (void) iscsit_config_merge(cfg);
2979 2979
2980 2980 it_config_free_cmn(cfg);
2981 2981
2982 2982 /* Restore old iSCSI driver global state */
2983 2983 mutex_enter(&iscsit_global.global_state_mutex);
2984 2984 ASSERT(iscsit_global.global_svc_state == ISE_BUSY ||
2985 2985 iscsit_global.global_svc_state == ISE_ENABLING);
2986 2986 iscsit_global.global_svc_state = old_state;
2987 2987 mutex_exit(&iscsit_global.global_state_mutex);
2988 2988 }
2989 2989
2990 2990
2991 2991 static it_cfg_status_t
2992 2992 iscsit_config_merge(it_config_t *in_cfg)
2993 2993 {
2994 2994 it_cfg_status_t status;
2995 2995 it_config_t *cfg;
2996 2996 it_config_t tmp_cfg;
2997 2997 list_t tpg_del_list;
2998 2998
2999 2999 if (in_cfg) {
3000 3000 cfg = in_cfg;
3001 3001 } else {
3002 3002 /* Make empty config */
3003 3003 bzero(&tmp_cfg, sizeof (tmp_cfg));
3004 3004 cfg = &tmp_cfg;
3005 3005 }
3006 3006
3007 3007 list_create(&tpg_del_list, sizeof (iscsit_tpg_t),
3008 3008 offsetof(iscsit_tpg_t, tpg_delete_ln));
3009 3009
3010 3010 /*
3011 3011 * Update targets, initiator contexts, target portal groups,
3012 3012 * and iSNS client
3013 3013 */
3014 3014 ISCSIT_GLOBAL_LOCK(RW_WRITER);
3015 3015 if (((status = iscsit_config_merge_tpg(cfg, &tpg_del_list))
3016 3016 != 0) ||
3017 3017 ((status = iscsit_config_merge_tgt(cfg)) != 0) ||
3018 3018 ((status = iscsit_config_merge_ini(cfg)) != 0) ||
3019 3019 ((status = isnst_config_merge(cfg)) != 0)) {
3020 3020 ISCSIT_GLOBAL_UNLOCK();
3021 3021 return (status);
3022 3022 }
3023 3023
3024 3024 /* Update other global config parameters */
3025 3025 if (iscsit_global.global_props) {
3026 3026 nvlist_free(iscsit_global.global_props);
3027 3027 iscsit_global.global_props = NULL;
3028 3028 }
3029 3029 if (in_cfg) {
3030 3030 (void) nvlist_dup(cfg->config_global_properties,
3031 3031 &iscsit_global.global_props, KM_SLEEP);
3032 3032 }
3033 3033 ISCSIT_GLOBAL_UNLOCK();
3034 3034
3035 3035 iscsit_config_destroy_tpgs(&tpg_del_list);
3036 3036
3037 3037 list_destroy(&tpg_del_list);
3038 3038
3039 3039 return (ITCFG_SUCCESS);
3040 3040 }
3041 3041
3042 3042 /*
3043 3043 * iscsit_sna_lt[e]
3044 3044 *
3045 3045 * Compare serial numbers using serial number arithmetic as defined in
3046 3046 * RFC 1982.
3047 3047 *
3048 3048 * NOTE: This code is duplicated in the isns server. It ought to be common.
3049 3049 */
3050 3050
3051 3051 static int
3052 3052 iscsit_sna_lt(uint32_t sn1, uint32_t sn2)
3053 3053 {
3054 3054 return ((sn1 != sn2) &&
3055 3055 (((sn1 < sn2) && ((sn2 - sn1) < ISCSIT_SNA32_CHECK)) ||
3056 3056 ((sn1 > sn2) && ((sn1 - sn2) > ISCSIT_SNA32_CHECK))));
3057 3057 }
3058 3058
3059 3059 static int
3060 3060 iscsit_sna_lte(uint32_t sn1, uint32_t sn2)
3061 3061 {
3062 3062 return ((sn1 == sn2) ||
3063 3063 (((sn1 < sn2) && ((sn2 - sn1) < ISCSIT_SNA32_CHECK)) ||
3064 3064 ((sn1 > sn2) && ((sn1 - sn2) > ISCSIT_SNA32_CHECK))));
3065 3065 }
3066 3066
3067 3067
3068 3068 static boolean_t
3069 3069 iscsit_cmdsn_in_window(iscsit_conn_t *ict, uint32_t cmdsn)
3070 3070 {
3071 3071 iscsit_sess_t *ist = ict->ict_sess;
3072 3072 int rval = B_TRUE;
3073 3073
3074 3074 ist = ict->ict_sess;
3075 3075
3076 3076 mutex_enter(&ist->ist_sn_mutex);
3077 3077
3078 3078 /*
3079 3079 * If cmdsn is less than ist_expcmdsn - iscsit_cmd_window() or
3080 3080 * greater than ist_expcmdsn, it's not in the window.
3081 3081 */
3082 3082
3083 3083 if (iscsit_sna_lt(cmdsn, (ist->ist_expcmdsn - iscsit_cmd_window())) ||
3084 3084 !iscsit_sna_lte(cmdsn, ist->ist_expcmdsn)) {
3085 3085 rval = B_FALSE;
3086 3086 }
3087 3087
3088 3088 mutex_exit(&ist->ist_sn_mutex);
3089 3089
3090 3090 return (rval);
3091 3091 }
3092 3092
3093 3093 /*
3094 3094 * iscsit_check_cmdsn_and_queue
3095 3095 *
3096 3096 * Independent of the order in which the iSCSI target receives non-immediate
3097 3097 * command PDU across the entire session and any multiple connections within
3098 3098 * the session, the target must deliver the commands to the SCSI layer in
3099 3099 * CmdSN order. So out-of-order non-immediate commands are queued up on a
3100 3100 * session-wide wait queue. Duplicate commands are ignored.
3101 3101 *
3102 3102 */
3103 3103 static int
3104 3104 iscsit_check_cmdsn_and_queue(idm_pdu_t *rx_pdu)
3105 3105 {
3106 3106 idm_conn_t *ic = rx_pdu->isp_ic;
3107 3107 iscsit_conn_t *ict = ic->ic_handle;
3108 3108 iscsit_sess_t *ist = ict->ict_sess;
3109 3109 iscsi_scsi_cmd_hdr_t *hdr = (iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr;
3110 3110
3111 3111 mutex_enter(&ist->ist_sn_mutex);
3112 3112 if (hdr->opcode & ISCSI_OP_IMMEDIATE) {
3113 3113 /* do not queue, handle it immediately */
3114 3114 DTRACE_PROBE2(immediate__cmd, iscsit_sess_t *, ist,
3115 3115 idm_pdu_t *, rx_pdu);
3116 3116 mutex_exit(&ist->ist_sn_mutex);
3117 3117 return (ISCSIT_CMDSN_EQ_EXPCMDSN);
3118 3118 }
3119 3119 if (iscsit_sna_lt(ist->ist_expcmdsn, ntohl(hdr->cmdsn))) {
3120 3120 /*
3121 3121 * Out-of-order commands (cmdSN higher than ExpCmdSN)
3122 3122 * are staged on a fixed-size circular buffer until
3123 3123 * the missing command is delivered to the SCSI layer.
3124 3124 * Irrespective of the order of insertion into the
3125 3125 * staging queue, the commands are processed out of the
3126 3126 * queue in cmdSN order only.
3127 3127 */
3128 3128 rx_pdu->isp_queue_time = gethrtime();
3129 3129 iscsit_add_pdu_to_queue(ist, rx_pdu);
3130 3130 mutex_exit(&ist->ist_sn_mutex);
3131 3131 return (ISCSIT_CMDSN_GT_EXPCMDSN);
3132 3132 } else if (iscsit_sna_lt(ntohl(hdr->cmdsn), ist->ist_expcmdsn)) {
3133 3133 DTRACE_PROBE3(cmdsn__lt__expcmdsn, iscsit_sess_t *, ist,
3134 3134 iscsit_conn_t *, ict, idm_pdu_t *, rx_pdu);
3135 3135 mutex_exit(&ist->ist_sn_mutex);
3136 3136 return (ISCSIT_CMDSN_LT_EXPCMDSN);
3137 3137 } else {
3138 3138 mutex_exit(&ist->ist_sn_mutex);
3139 3139 return (ISCSIT_CMDSN_EQ_EXPCMDSN);
3140 3140 }
3141 3141 }
3142 3142
3143 3143 /*
3144 3144 * iscsit_add_pdu_to_queue() adds PDUs into the array indexed by
3145 3145 * their cmdsn value. The length of the array is kept above the
3146 3146 * maximum window size. The window keeps the cmdsn within a range
3147 3147 * such that there are no collisons. e.g. the assumption is that
3148 3148 * the windowing checks make it impossible to receive PDUs that
3149 3149 * index into the same location in the array.
3150 3150 */
3151 3151 static void
3152 3152 iscsit_add_pdu_to_queue(iscsit_sess_t *ist, idm_pdu_t *rx_pdu)
3153 3153 {
3154 3154 iscsit_cbuf_t *cbuf = ist->ist_rxpdu_queue;
3155 3155 iscsit_conn_t *ict = rx_pdu->isp_ic->ic_handle;
3156 3156 uint32_t cmdsn =
3157 3157 ((iscsi_scsi_cmd_hdr_t *)rx_pdu->isp_hdr)->cmdsn;
3158 3158 uint32_t index;
3159 3159
3160 3160 ASSERT(MUTEX_HELD(&ist->ist_sn_mutex));
3161 3161 /*
3162 3162 * If the connection is being torn down, then
3163 3163 * don't add the PDU to the staging queue
3164 3164 */
3165 3165 mutex_enter(&ict->ict_mutex);
3166 3166 if (ict->ict_lost) {
3167 3167 mutex_exit(&ict->ict_mutex);
3168 3168 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
3169 3169 return;
3170 3170 }
3171 3171 iscsit_conn_dispatch_hold(ict);
3172 3172 mutex_exit(&ict->ict_mutex);
3173 3173
3174 3174 index = ntohl(cmdsn) % ISCSIT_RXPDU_QUEUE_LEN;
3175 3175 /*
3176 3176 * In the normal case, assuming that the Initiator is not
3177 3177 * buggy and that we don't have packet duplication occuring,
3178 3178 * the entry in the array will be NULL. However, we may have
3179 3179 * received a duplicate PDU with cmdsn > expsn , and in that
3180 3180 * case we just ignore this PDU -- the previously received one
3181 3181 * remains queued for processing. We need to be careful not
3182 3182 * to leak this one however.
3183 3183 */
3184 3184 if (cbuf->cb_buffer[index] != NULL) {
3185 3185 idm_pdu_complete(rx_pdu, IDM_STATUS_FAIL);
3186 3186 } else {
3187 3187 cbuf->cb_buffer[index] = rx_pdu;
3188 3188 cbuf->cb_num_elems++;
3189 3189 }
3190 3190 }
3191 3191
3192 3192 static idm_pdu_t *
3193 3193 iscsit_remove_pdu_from_queue(iscsit_sess_t *ist, uint32_t cmdsn)
3194 3194 {
3195 3195 iscsit_cbuf_t *cbuf = ist->ist_rxpdu_queue;
3196 3196 idm_pdu_t *pdu = NULL;
3197 3197 uint32_t index;
3198 3198
3199 3199 ASSERT(MUTEX_HELD(&ist->ist_sn_mutex));
3200 3200 index = cmdsn % ISCSIT_RXPDU_QUEUE_LEN;
3201 3201 if ((pdu = cbuf->cb_buffer[index]) != NULL) {
3202 3202 ASSERT(cmdsn ==
3203 3203 ntohl(((iscsi_scsi_cmd_hdr_t *)pdu->isp_hdr)->cmdsn));
3204 3204 cbuf->cb_buffer[index] = NULL;
3205 3205 cbuf->cb_num_elems--;
3206 3206 return (pdu);
3207 3207 }
3208 3208 return (NULL);
3209 3209 }
3210 3210
3211 3211 /*
3212 3212 * iscsit_process_pdu_in_queue() finds the next pdu in sequence
3213 3213 * and posts it to the SCSI layer
3214 3214 */
3215 3215 static void
3216 3216 iscsit_process_pdu_in_queue(iscsit_sess_t *ist)
3217 3217 {
3218 3218 iscsit_cbuf_t *cbuf = ist->ist_rxpdu_queue;
3219 3219 idm_pdu_t *pdu = NULL;
3220 3220 uint32_t expcmdsn;
3221 3221
3222 3222 for (;;) {
3223 3223 mutex_enter(&ist->ist_sn_mutex);
3224 3224 if (cbuf->cb_num_elems == 0) {
3225 3225 mutex_exit(&ist->ist_sn_mutex);
3226 3226 break;
3227 3227 }
3228 3228 expcmdsn = ist->ist_expcmdsn;
3229 3229 if ((pdu = iscsit_remove_pdu_from_queue(ist, expcmdsn))
3230 3230 == NULL) {
3231 3231 mutex_exit(&ist->ist_sn_mutex);
3232 3232 break;
3233 3233 }
3234 3234 mutex_exit(&ist->ist_sn_mutex);
3235 3235 iscsit_post_staged_pdu(pdu);
3236 3236 }
3237 3237 }
3238 3238
3239 3239 static void
3240 3240 iscsit_post_staged_pdu(idm_pdu_t *rx_pdu)
3241 3241 {
3242 3242 iscsit_conn_t *ict = rx_pdu->isp_ic->ic_handle;
3243 3243
3244 3244 /* Post the PDU to the SCSI layer */
3245 3245 switch (IDM_PDU_OPCODE(rx_pdu)) {
3246 3246 case ISCSI_OP_NOOP_OUT:
3247 3247 iscsit_set_cmdsn(ict, rx_pdu);
3248 3248 iscsit_pdu_op_noop(ict, rx_pdu);
3249 3249 break;
3250 3250 case ISCSI_OP_TEXT_CMD:
3251 3251 iscsit_set_cmdsn(ict, rx_pdu);
3252 3252 iscsit_pdu_op_text_cmd(ict, rx_pdu);
3253 3253 break;
3254 3254 case ISCSI_OP_SCSI_TASK_MGT_MSG:
3255 3255 iscsit_set_cmdsn(ict, rx_pdu);
3256 3256 iscsit_op_scsi_task_mgmt(ict, rx_pdu);
3257 3257 break;
3258 3258 case ISCSI_OP_SCSI_CMD:
3259 3259 /* cmdSN will be incremented after creating itask */
3260 3260 iscsit_post_scsi_cmd(rx_pdu->isp_ic, rx_pdu);
3261 3261 break;
3262 3262 case ISCSI_OP_LOGOUT_CMD:
3263 3263 iscsit_set_cmdsn(ict, rx_pdu);
3264 3264 iscsit_pdu_op_logout_cmd(ict, rx_pdu);
3265 3265 break;
3266 3266 default:
3267 3267 /* No other PDUs should be placed on the queue */
3268 3268 ASSERT(0);
3269 3269 }
3270 3270 iscsit_conn_dispatch_rele(ict); /* release hold on the conn */
3271 3271 }
3272 3272
3273 3273 /* ARGSUSED */
3274 3274 void
3275 3275 iscsit_rxpdu_queue_monitor_start(void)
3276 3276 {
3277 3277 mutex_enter(&iscsit_rxpdu_queue_monitor_mutex);
3278 3278 if (iscsit_rxpdu_queue_monitor_thr_running) {
3279 3279 mutex_exit(&iscsit_rxpdu_queue_monitor_mutex);
3280 3280 return;
3281 3281 }
3282 3282 iscsit_rxpdu_queue_monitor_thr_id =
3283 3283 thread_create(NULL, 0, iscsit_rxpdu_queue_monitor, NULL,
3284 3284 0, &p0, TS_RUN, minclsyspri);
3285 3285 while (!iscsit_rxpdu_queue_monitor_thr_running) {
3286 3286 cv_wait(&iscsit_rxpdu_queue_monitor_cv,
3287 3287 &iscsit_rxpdu_queue_monitor_mutex);
3288 3288 }
3289 3289 mutex_exit(&iscsit_rxpdu_queue_monitor_mutex);
3290 3290
3291 3291 }
3292 3292
3293 3293 /* ARGSUSED */
3294 3294 void
3295 3295 iscsit_rxpdu_queue_monitor_stop(void)
3296 3296 {
3297 3297 mutex_enter(&iscsit_rxpdu_queue_monitor_mutex);
3298 3298 if (iscsit_rxpdu_queue_monitor_thr_running) {
3299 3299 iscsit_rxpdu_queue_monitor_thr_running = B_FALSE;
3300 3300 cv_signal(&iscsit_rxpdu_queue_monitor_cv);
3301 3301 mutex_exit(&iscsit_rxpdu_queue_monitor_mutex);
3302 3302
3303 3303 thread_join(iscsit_rxpdu_queue_monitor_thr_did);
3304 3304 return;
3305 3305 }
3306 3306 mutex_exit(&iscsit_rxpdu_queue_monitor_mutex);
3307 3307 }
3308 3308
3309 3309 /*
3310 3310 * A separate thread is used to scan the staging queue on all the
3311 3311 * sessions, If a delayed PDU does not arrive within a timeout, the
3312 3312 * target will advance to the staged PDU that is next in sequence
3313 3313 * and exceeded the threshold wait time. It is up to the initiator
3314 3314 * to note that the target has not acknowledged a particular cmdsn
3315 3315 * and take appropriate action.
3316 3316 */
3317 3317 /* ARGSUSED */
3318 3318 static void
3319 3319 iscsit_rxpdu_queue_monitor(void *arg)
3320 3320 {
3321 3321 iscsit_tgt_t *tgt;
3322 3322 iscsit_sess_t *ist;
3323 3323
3324 3324 mutex_enter(&iscsit_rxpdu_queue_monitor_mutex);
3325 3325 iscsit_rxpdu_queue_monitor_thr_did = curthread->t_did;
3326 3326 iscsit_rxpdu_queue_monitor_thr_running = B_TRUE;
3327 3327 cv_signal(&iscsit_rxpdu_queue_monitor_cv);
3328 3328
3329 3329 while (iscsit_rxpdu_queue_monitor_thr_running) {
3330 3330 ISCSIT_GLOBAL_LOCK(RW_READER);
3331 3331 for (tgt = avl_first(&iscsit_global.global_target_list);
3332 3332 tgt != NULL;
3333 3333 tgt = AVL_NEXT(&iscsit_global.global_target_list, tgt)) {
3334 3334 mutex_enter(&tgt->target_mutex);
3335 3335 for (ist = avl_first(&tgt->target_sess_list);
3336 3336 ist != NULL;
3337 3337 ist = AVL_NEXT(&tgt->target_sess_list, ist)) {
3338 3338
↓ open down ↓ |
3338 lines elided |
↑ open up ↑ |
3339 3339 iscsit_rxpdu_queue_monitor_session(ist);
3340 3340 }
3341 3341 mutex_exit(&tgt->target_mutex);
3342 3342 }
3343 3343 ISCSIT_GLOBAL_UNLOCK();
3344 3344 if (iscsit_rxpdu_queue_monitor_thr_running == B_FALSE) {
3345 3345 break;
3346 3346 }
3347 3347 (void) cv_reltimedwait(&iscsit_rxpdu_queue_monitor_cv,
3348 3348 &iscsit_rxpdu_queue_monitor_mutex,
3349 - ISCSIT_RXPDU_QUEUE_MONITOR_INTERVAL * drv_usectohz(1000000),
3349 + drv_sectohz(ISCSIT_RXPDU_QUEUE_MONITOR_INTERVAL),
3350 3350 TR_CLOCK_TICK);
3351 3351 }
3352 3352 mutex_exit(&iscsit_rxpdu_queue_monitor_mutex);
3353 3353 thread_exit();
3354 3354 }
3355 3355
3356 3356 static void
3357 3357 iscsit_rxpdu_queue_monitor_session(iscsit_sess_t *ist)
3358 3358 {
3359 3359 iscsit_cbuf_t *cbuf = ist->ist_rxpdu_queue;
3360 3360 idm_pdu_t *next_pdu = NULL;
3361 3361 uint32_t index, next_cmdsn, i;
3362 3362
3363 3363 /*
3364 3364 * Assume that all PDUs in the staging queue have a cmdsn >= expcmdsn.
3365 3365 * Starting with the expcmdsn, iterate over the staged PDUs to find
3366 3366 * the next PDU with a wait time greater than the threshold. If found
3367 3367 * advance the staged PDU to the SCSI layer, skipping over the missing
3368 3368 * PDU(s) to get past the hole in the command sequence. It is up to
3369 3369 * the initiator to note that the target has not acknowledged a cmdsn
3370 3370 * and take appropriate action.
3371 3371 *
3372 3372 * Since the PDU(s) arrive in any random order, it is possible that
3373 3373 * that the actual wait time for a particular PDU is much longer than
3374 3374 * the defined threshold. e.g. Consider a case where commands are sent
3375 3375 * over 4 different connections, and cmdsn = 1004 arrives first, then
3376 3376 * 1003, and 1002 and 1001 are lost due to a connection failure.
3377 3377 * So now 1003 is waiting for 1002 to be delivered, and although the
3378 3378 * wait time of 1004 > wait time of 1003, only 1003 will be considered
3379 3379 * by the monitor thread. 1004 will be automatically processed by
3380 3380 * iscsit_process_pdu_in_queue() once the scan is complete and the
3381 3381 * expcmdsn becomes current.
3382 3382 */
3383 3383 mutex_enter(&ist->ist_sn_mutex);
3384 3384 cbuf = ist->ist_rxpdu_queue;
3385 3385 if (cbuf->cb_num_elems == 0) {
3386 3386 mutex_exit(&ist->ist_sn_mutex);
3387 3387 return;
3388 3388 }
3389 3389 for (next_pdu = NULL, i = 0; ; i++) {
3390 3390 next_cmdsn = ist->ist_expcmdsn + i; /* start at expcmdsn */
3391 3391 index = next_cmdsn % ISCSIT_RXPDU_QUEUE_LEN;
3392 3392 if ((next_pdu = cbuf->cb_buffer[index]) != NULL) {
3393 3393 /*
3394 3394 * If the PDU wait time has not exceeded threshold
3395 3395 * stop scanning the staging queue until the timer
3396 3396 * fires again
3397 3397 */
3398 3398 if ((gethrtime() - next_pdu->isp_queue_time)
3399 3399 < (rxpdu_queue_threshold * NANOSEC)) {
3400 3400 mutex_exit(&ist->ist_sn_mutex);
3401 3401 return;
3402 3402 }
3403 3403 /*
3404 3404 * Remove the next PDU from the queue and post it
3405 3405 * to the SCSI layer, skipping over the missing
3406 3406 * PDU. Stop scanning the staging queue until
3407 3407 * the monitor timer fires again
3408 3408 */
3409 3409 (void) iscsit_remove_pdu_from_queue(ist, next_cmdsn);
3410 3410 mutex_exit(&ist->ist_sn_mutex);
3411 3411 DTRACE_PROBE3(advanced__to__blocked__cmdsn,
3412 3412 iscsit_sess_t *, ist, idm_pdu_t *, next_pdu,
3413 3413 uint32_t, next_cmdsn);
3414 3414 iscsit_post_staged_pdu(next_pdu);
3415 3415 /* Deliver any subsequent PDUs immediately */
3416 3416 iscsit_process_pdu_in_queue(ist);
3417 3417 return;
3418 3418 }
3419 3419 /*
3420 3420 * Skipping over i PDUs, e.g. a case where commands 1001 and
3421 3421 * 1002 are lost in the network, skip over both and post 1003
3422 3422 * expcmdsn then becomes 1004 at the end of the scan.
3423 3423 */
3424 3424 DTRACE_PROBE2(skipping__over__cmdsn, iscsit_sess_t *, ist,
3425 3425 uint32_t, next_cmdsn);
3426 3426 }
3427 3427 /*
3428 3428 * following the assumption, staged cmdsn >= expcmdsn, this statement
3429 3429 * is never reached.
3430 3430 */
3431 3431 }
↓ open down ↓ |
72 lines elided |
↑ open up ↑ |
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX