1 | // SPDX-License-Identifier: GPL-2.0-or-later |
2 | /* |
3 | * iSCSI Initiator over TCP/IP Data-Path |
4 | * |
5 | * Copyright (C) 2004 Dmitry Yusupov |
6 | * Copyright (C) 2004 Alex Aizman |
7 | * Copyright (C) 2005 - 2006 Mike Christie |
8 | * Copyright (C) 2006 Red Hat, Inc. All rights reserved. |
9 | * maintained by open-iscsi@googlegroups.com |
10 | * |
11 | * See the file COPYING included with this distribution for more details. |
12 | * |
13 | * Credits: |
14 | * Christoph Hellwig |
15 | * FUJITA Tomonori |
16 | * Arne Redlich |
17 | * Zhenyu Wang |
18 | */ |
19 | |
20 | #include <crypto/hash.h> |
21 | #include <linux/types.h> |
22 | #include <linux/inet.h> |
23 | #include <linux/slab.h> |
24 | #include <linux/sched/mm.h> |
25 | #include <linux/file.h> |
26 | #include <linux/blkdev.h> |
27 | #include <linux/delay.h> |
28 | #include <linux/kfifo.h> |
29 | #include <linux/scatterlist.h> |
30 | #include <linux/module.h> |
31 | #include <linux/backing-dev.h> |
32 | #include <net/tcp.h> |
33 | #include <scsi/scsi_cmnd.h> |
34 | #include <scsi/scsi_device.h> |
35 | #include <scsi/scsi_host.h> |
36 | #include <scsi/scsi.h> |
37 | #include <scsi/scsi_transport_iscsi.h> |
38 | #include <trace/events/iscsi.h> |
39 | #include <trace/events/sock.h> |
40 | |
41 | #include "iscsi_tcp.h" |
42 | |
43 | MODULE_AUTHOR("Mike Christie <michaelc@cs.wisc.edu>, " |
44 | "Dmitry Yusupov <dmitry_yus@yahoo.com>, " |
45 | "Alex Aizman <itn780@yahoo.com>" ); |
46 | MODULE_DESCRIPTION("iSCSI/TCP data-path" ); |
47 | MODULE_LICENSE("GPL" ); |
48 | |
49 | static struct scsi_transport_template *iscsi_sw_tcp_scsi_transport; |
50 | static const struct scsi_host_template iscsi_sw_tcp_sht; |
51 | static struct iscsi_transport iscsi_sw_tcp_transport; |
52 | |
53 | static unsigned int iscsi_max_lun = ~0; |
54 | module_param_named(max_lun, iscsi_max_lun, uint, S_IRUGO); |
55 | |
56 | static bool iscsi_recv_from_iscsi_q; |
57 | module_param_named(recv_from_iscsi_q, iscsi_recv_from_iscsi_q, bool, 0644); |
58 | MODULE_PARM_DESC(recv_from_iscsi_q, "Set to true to read iSCSI data/headers from the iscsi_q workqueue. The default is false which will perform reads from the network softirq context." ); |
59 | |
60 | static int iscsi_sw_tcp_dbg; |
61 | module_param_named(debug_iscsi_tcp, iscsi_sw_tcp_dbg, int, |
62 | S_IRUGO | S_IWUSR); |
63 | MODULE_PARM_DESC(debug_iscsi_tcp, "Turn on debugging for iscsi_tcp module " |
64 | "Set to 1 to turn on, and zero to turn off. Default is off." ); |
65 | |
66 | #define ISCSI_SW_TCP_DBG(_conn, dbg_fmt, arg...) \ |
67 | do { \ |
68 | if (iscsi_sw_tcp_dbg) \ |
69 | iscsi_conn_printk(KERN_INFO, _conn, \ |
70 | "%s " dbg_fmt, \ |
71 | __func__, ##arg); \ |
72 | iscsi_dbg_trace(trace_iscsi_dbg_sw_tcp, \ |
73 | &(_conn)->cls_conn->dev, \ |
74 | "%s " dbg_fmt, __func__, ##arg);\ |
75 | } while (0); |
76 | |
77 | |
78 | /** |
79 | * iscsi_sw_tcp_recv - TCP receive in sendfile fashion |
80 | * @rd_desc: read descriptor |
81 | * @skb: socket buffer |
82 | * @offset: offset in skb |
83 | * @len: skb->len - offset |
84 | */ |
85 | static int iscsi_sw_tcp_recv(read_descriptor_t *rd_desc, struct sk_buff *skb, |
86 | unsigned int offset, size_t len) |
87 | { |
88 | struct iscsi_conn *conn = rd_desc->arg.data; |
89 | unsigned int consumed, total_consumed = 0; |
90 | int status; |
91 | |
92 | ISCSI_SW_TCP_DBG(conn, "in %d bytes\n" , skb->len - offset); |
93 | |
94 | do { |
95 | status = 0; |
96 | consumed = iscsi_tcp_recv_skb(conn, skb, offset, offloaded: 0, status: &status); |
97 | offset += consumed; |
98 | total_consumed += consumed; |
99 | } while (consumed != 0 && status != ISCSI_TCP_SKB_DONE); |
100 | |
101 | ISCSI_SW_TCP_DBG(conn, "read %d bytes status %d\n" , |
102 | skb->len - offset, status); |
103 | return total_consumed; |
104 | } |
105 | |
106 | /** |
107 | * iscsi_sw_sk_state_check - check socket state |
108 | * @sk: socket |
109 | * |
110 | * If the socket is in CLOSE or CLOSE_WAIT we should |
111 | * not close the connection if there is still some |
112 | * data pending. |
113 | * |
114 | * Must be called with sk_callback_lock. |
115 | */ |
116 | static inline int iscsi_sw_sk_state_check(struct sock *sk) |
117 | { |
118 | struct iscsi_conn *conn = sk->sk_user_data; |
119 | |
120 | if ((sk->sk_state == TCP_CLOSE_WAIT || sk->sk_state == TCP_CLOSE) && |
121 | (conn->session->state != ISCSI_STATE_LOGGING_OUT) && |
122 | !atomic_read(v: &sk->sk_rmem_alloc)) { |
123 | ISCSI_SW_TCP_DBG(conn, "TCP_CLOSE|TCP_CLOSE_WAIT\n" ); |
124 | iscsi_conn_failure(conn, err: ISCSI_ERR_TCP_CONN_CLOSE); |
125 | return -ECONNRESET; |
126 | } |
127 | return 0; |
128 | } |
129 | |
130 | static void iscsi_sw_tcp_recv_data(struct iscsi_conn *conn) |
131 | { |
132 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
133 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
134 | struct sock *sk = tcp_sw_conn->sock->sk; |
135 | read_descriptor_t rd_desc; |
136 | |
137 | /* |
138 | * Use rd_desc to pass 'conn' to iscsi_tcp_recv. |
139 | * We set count to 1 because we want the network layer to |
140 | * hand us all the skbs that are available. iscsi_tcp_recv |
141 | * handled pdus that cross buffers or pdus that still need data. |
142 | */ |
143 | rd_desc.arg.data = conn; |
144 | rd_desc.count = 1; |
145 | |
146 | tcp_read_sock(sk, desc: &rd_desc, recv_actor: iscsi_sw_tcp_recv); |
147 | |
148 | /* If we had to (atomically) map a highmem page, |
149 | * unmap it now. */ |
150 | iscsi_tcp_segment_unmap(segment: &tcp_conn->in.segment); |
151 | |
152 | iscsi_sw_sk_state_check(sk); |
153 | } |
154 | |
155 | static void iscsi_sw_tcp_recv_data_work(struct work_struct *work) |
156 | { |
157 | struct iscsi_conn *conn = container_of(work, struct iscsi_conn, |
158 | recvwork); |
159 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
160 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
161 | struct sock *sk = tcp_sw_conn->sock->sk; |
162 | |
163 | lock_sock(sk); |
164 | iscsi_sw_tcp_recv_data(conn); |
165 | release_sock(sk); |
166 | } |
167 | |
168 | static void iscsi_sw_tcp_data_ready(struct sock *sk) |
169 | { |
170 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
171 | struct iscsi_tcp_conn *tcp_conn; |
172 | struct iscsi_conn *conn; |
173 | |
174 | trace_sk_data_ready(sk); |
175 | |
176 | read_lock_bh(&sk->sk_callback_lock); |
177 | conn = sk->sk_user_data; |
178 | if (!conn) { |
179 | read_unlock_bh(&sk->sk_callback_lock); |
180 | return; |
181 | } |
182 | tcp_conn = conn->dd_data; |
183 | tcp_sw_conn = tcp_conn->dd_data; |
184 | |
185 | if (tcp_sw_conn->queue_recv) |
186 | iscsi_conn_queue_recv(conn); |
187 | else |
188 | iscsi_sw_tcp_recv_data(conn); |
189 | read_unlock_bh(&sk->sk_callback_lock); |
190 | } |
191 | |
192 | static void iscsi_sw_tcp_state_change(struct sock *sk) |
193 | { |
194 | struct iscsi_tcp_conn *tcp_conn; |
195 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
196 | struct iscsi_conn *conn; |
197 | void (*old_state_change)(struct sock *); |
198 | |
199 | read_lock_bh(&sk->sk_callback_lock); |
200 | conn = sk->sk_user_data; |
201 | if (!conn) { |
202 | read_unlock_bh(&sk->sk_callback_lock); |
203 | return; |
204 | } |
205 | |
206 | iscsi_sw_sk_state_check(sk); |
207 | |
208 | tcp_conn = conn->dd_data; |
209 | tcp_sw_conn = tcp_conn->dd_data; |
210 | old_state_change = tcp_sw_conn->old_state_change; |
211 | |
212 | read_unlock_bh(&sk->sk_callback_lock); |
213 | |
214 | old_state_change(sk); |
215 | } |
216 | |
217 | /** |
218 | * iscsi_sw_tcp_write_space - Called when more output buffer space is available |
219 | * @sk: socket space is available for |
220 | **/ |
221 | static void iscsi_sw_tcp_write_space(struct sock *sk) |
222 | { |
223 | struct iscsi_conn *conn; |
224 | struct iscsi_tcp_conn *tcp_conn; |
225 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
226 | void (*old_write_space)(struct sock *); |
227 | |
228 | read_lock_bh(&sk->sk_callback_lock); |
229 | conn = sk->sk_user_data; |
230 | if (!conn) { |
231 | read_unlock_bh(&sk->sk_callback_lock); |
232 | return; |
233 | } |
234 | |
235 | tcp_conn = conn->dd_data; |
236 | tcp_sw_conn = tcp_conn->dd_data; |
237 | old_write_space = tcp_sw_conn->old_write_space; |
238 | read_unlock_bh(&sk->sk_callback_lock); |
239 | |
240 | old_write_space(sk); |
241 | |
242 | ISCSI_SW_TCP_DBG(conn, "iscsi_write_space\n" ); |
243 | iscsi_conn_queue_xmit(conn); |
244 | } |
245 | |
246 | static void iscsi_sw_tcp_conn_set_callbacks(struct iscsi_conn *conn) |
247 | { |
248 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
249 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
250 | struct sock *sk = tcp_sw_conn->sock->sk; |
251 | |
252 | /* assign new callbacks */ |
253 | write_lock_bh(&sk->sk_callback_lock); |
254 | sk->sk_user_data = conn; |
255 | tcp_sw_conn->old_data_ready = sk->sk_data_ready; |
256 | tcp_sw_conn->old_state_change = sk->sk_state_change; |
257 | tcp_sw_conn->old_write_space = sk->sk_write_space; |
258 | sk->sk_data_ready = iscsi_sw_tcp_data_ready; |
259 | sk->sk_state_change = iscsi_sw_tcp_state_change; |
260 | sk->sk_write_space = iscsi_sw_tcp_write_space; |
261 | write_unlock_bh(&sk->sk_callback_lock); |
262 | } |
263 | |
264 | static void |
265 | iscsi_sw_tcp_conn_restore_callbacks(struct iscsi_conn *conn) |
266 | { |
267 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
268 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
269 | struct sock *sk = tcp_sw_conn->sock->sk; |
270 | |
271 | /* restore socket callbacks, see also: iscsi_conn_set_callbacks() */ |
272 | write_lock_bh(&sk->sk_callback_lock); |
273 | sk->sk_user_data = NULL; |
274 | sk->sk_data_ready = tcp_sw_conn->old_data_ready; |
275 | sk->sk_state_change = tcp_sw_conn->old_state_change; |
276 | sk->sk_write_space = tcp_sw_conn->old_write_space; |
277 | sk->sk_no_check_tx = 0; |
278 | write_unlock_bh(&sk->sk_callback_lock); |
279 | } |
280 | |
281 | /** |
282 | * iscsi_sw_tcp_xmit_segment - transmit segment |
283 | * @tcp_conn: the iSCSI TCP connection |
284 | * @segment: the buffer to transmnit |
285 | * |
286 | * This function transmits as much of the buffer as |
287 | * the network layer will accept, and returns the number of |
288 | * bytes transmitted. |
289 | * |
290 | * If CRC hashing is enabled, the function will compute the |
291 | * hash as it goes. When the entire segment has been transmitted, |
292 | * it will retrieve the hash value and send it as well. |
293 | */ |
294 | static int iscsi_sw_tcp_xmit_segment(struct iscsi_tcp_conn *tcp_conn, |
295 | struct iscsi_segment *segment) |
296 | { |
297 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
298 | struct socket *sk = tcp_sw_conn->sock; |
299 | unsigned int copied = 0; |
300 | int r = 0; |
301 | |
302 | while (!iscsi_tcp_segment_done(tcp_conn, segment, recv: 0, copied: r)) { |
303 | struct scatterlist *sg; |
304 | struct msghdr msg = {}; |
305 | struct bio_vec bv; |
306 | unsigned int offset, copy; |
307 | |
308 | r = 0; |
309 | offset = segment->copied; |
310 | copy = segment->size - offset; |
311 | |
312 | if (segment->total_copied + segment->size < segment->total_size) |
313 | msg.msg_flags |= MSG_MORE; |
314 | |
315 | if (tcp_sw_conn->queue_recv) |
316 | msg.msg_flags |= MSG_DONTWAIT; |
317 | |
318 | if (!segment->data) { |
319 | if (!tcp_conn->iscsi_conn->datadgst_en) |
320 | msg.msg_flags |= MSG_SPLICE_PAGES; |
321 | sg = segment->sg; |
322 | offset += segment->sg_offset + sg->offset; |
323 | bvec_set_page(bv: &bv, page: sg_page(sg), len: copy, offset); |
324 | } else { |
325 | bvec_set_virt(bv: &bv, vaddr: segment->data + offset, len: copy); |
326 | } |
327 | iov_iter_bvec(i: &msg.msg_iter, ITER_SOURCE, bvec: &bv, nr_segs: 1, count: copy); |
328 | |
329 | r = sock_sendmsg(sock: sk, msg: &msg); |
330 | if (r < 0) { |
331 | iscsi_tcp_segment_unmap(segment); |
332 | return r; |
333 | } |
334 | copied += r; |
335 | } |
336 | return copied; |
337 | } |
338 | |
339 | /** |
340 | * iscsi_sw_tcp_xmit - TCP transmit |
341 | * @conn: iscsi connection |
342 | **/ |
343 | static int iscsi_sw_tcp_xmit(struct iscsi_conn *conn) |
344 | { |
345 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
346 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
347 | struct iscsi_segment *segment = &tcp_sw_conn->out.segment; |
348 | unsigned int consumed = 0; |
349 | int rc = 0; |
350 | |
351 | while (1) { |
352 | rc = iscsi_sw_tcp_xmit_segment(tcp_conn, segment); |
353 | /* |
354 | * We may not have been able to send data because the conn |
355 | * is getting stopped. libiscsi will know so propagate err |
356 | * for it to do the right thing. |
357 | */ |
358 | if (rc == -EAGAIN) |
359 | return rc; |
360 | else if (rc < 0) { |
361 | rc = ISCSI_ERR_XMIT_FAILED; |
362 | goto error; |
363 | } else if (rc == 0) |
364 | break; |
365 | |
366 | consumed += rc; |
367 | |
368 | if (segment->total_copied >= segment->total_size) { |
369 | if (segment->done != NULL) { |
370 | rc = segment->done(tcp_conn, segment); |
371 | if (rc != 0) |
372 | goto error; |
373 | } |
374 | } |
375 | } |
376 | |
377 | ISCSI_SW_TCP_DBG(conn, "xmit %d bytes\n" , consumed); |
378 | |
379 | conn->txdata_octets += consumed; |
380 | return consumed; |
381 | |
382 | error: |
383 | /* Transmit error. We could initiate error recovery |
384 | * here. */ |
385 | ISCSI_SW_TCP_DBG(conn, "Error sending PDU, errno=%d\n" , rc); |
386 | iscsi_conn_failure(conn, err: rc); |
387 | return -EIO; |
388 | } |
389 | |
390 | /** |
391 | * iscsi_sw_tcp_xmit_qlen - return the number of bytes queued for xmit |
392 | * @conn: iscsi connection |
393 | */ |
394 | static inline int iscsi_sw_tcp_xmit_qlen(struct iscsi_conn *conn) |
395 | { |
396 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
397 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
398 | struct iscsi_segment *segment = &tcp_sw_conn->out.segment; |
399 | |
400 | return segment->total_copied - segment->total_size; |
401 | } |
402 | |
403 | static int iscsi_sw_tcp_pdu_xmit(struct iscsi_task *task) |
404 | { |
405 | struct iscsi_conn *conn = task->conn; |
406 | unsigned int noreclaim_flag; |
407 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
408 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
409 | int rc = 0; |
410 | |
411 | if (!tcp_sw_conn->sock) { |
412 | iscsi_conn_printk(KERN_ERR, conn, |
413 | "Transport not bound to socket!\n" ); |
414 | return -EINVAL; |
415 | } |
416 | |
417 | noreclaim_flag = memalloc_noreclaim_save(); |
418 | |
419 | while (iscsi_sw_tcp_xmit_qlen(conn)) { |
420 | rc = iscsi_sw_tcp_xmit(conn); |
421 | if (rc == 0) { |
422 | rc = -EAGAIN; |
423 | break; |
424 | } |
425 | if (rc < 0) |
426 | break; |
427 | rc = 0; |
428 | } |
429 | |
430 | memalloc_noreclaim_restore(flags: noreclaim_flag); |
431 | return rc; |
432 | } |
433 | |
434 | /* |
435 | * This is called when we're done sending the header. |
436 | * Simply copy the data_segment to the send segment, and return. |
437 | */ |
438 | static int iscsi_sw_tcp_send_hdr_done(struct iscsi_tcp_conn *tcp_conn, |
439 | struct iscsi_segment *segment) |
440 | { |
441 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
442 | |
443 | tcp_sw_conn->out.segment = tcp_sw_conn->out.data_segment; |
444 | ISCSI_SW_TCP_DBG(tcp_conn->iscsi_conn, |
445 | "Header done. Next segment size %u total_size %u\n" , |
446 | tcp_sw_conn->out.segment.size, |
447 | tcp_sw_conn->out.segment.total_size); |
448 | return 0; |
449 | } |
450 | |
451 | static void iscsi_sw_tcp_send_hdr_prep(struct iscsi_conn *conn, void *hdr, |
452 | size_t hdrlen) |
453 | { |
454 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
455 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
456 | |
457 | ISCSI_SW_TCP_DBG(conn, "%s\n" , conn->hdrdgst_en ? |
458 | "digest enabled" : "digest disabled" ); |
459 | |
460 | /* Clear the data segment - needs to be filled in by the |
461 | * caller using iscsi_tcp_send_data_prep() */ |
462 | memset(&tcp_sw_conn->out.data_segment, 0, |
463 | sizeof(struct iscsi_segment)); |
464 | |
465 | /* If header digest is enabled, compute the CRC and |
466 | * place the digest into the same buffer. We make |
467 | * sure that both iscsi_tcp_task and mtask have |
468 | * sufficient room. |
469 | */ |
470 | if (conn->hdrdgst_en) { |
471 | iscsi_tcp_dgst_header(hash: tcp_sw_conn->tx_hash, hdr, hdrlen, |
472 | digest: hdr + hdrlen); |
473 | hdrlen += ISCSI_DIGEST_SIZE; |
474 | } |
475 | |
476 | /* Remember header pointer for later, when we need |
477 | * to decide whether there's a payload to go along |
478 | * with the header. */ |
479 | tcp_sw_conn->out.hdr = hdr; |
480 | |
481 | iscsi_segment_init_linear(segment: &tcp_sw_conn->out.segment, data: hdr, size: hdrlen, |
482 | done: iscsi_sw_tcp_send_hdr_done, NULL); |
483 | } |
484 | |
485 | /* |
486 | * Prepare the send buffer for the payload data. |
487 | * Padding and checksumming will all be taken care |
488 | * of by the iscsi_segment routines. |
489 | */ |
490 | static int |
491 | iscsi_sw_tcp_send_data_prep(struct iscsi_conn *conn, struct scatterlist *sg, |
492 | unsigned int count, unsigned int offset, |
493 | unsigned int len) |
494 | { |
495 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
496 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
497 | struct ahash_request *tx_hash = NULL; |
498 | unsigned int hdr_spec_len; |
499 | |
500 | ISCSI_SW_TCP_DBG(conn, "offset=%d, datalen=%d %s\n" , offset, len, |
501 | conn->datadgst_en ? |
502 | "digest enabled" : "digest disabled" ); |
503 | |
504 | /* Make sure the datalen matches what the caller |
505 | said he would send. */ |
506 | hdr_spec_len = ntoh24(tcp_sw_conn->out.hdr->dlength); |
507 | WARN_ON(iscsi_padded(len) != iscsi_padded(hdr_spec_len)); |
508 | |
509 | if (conn->datadgst_en) |
510 | tx_hash = tcp_sw_conn->tx_hash; |
511 | |
512 | return iscsi_segment_seek_sg(segment: &tcp_sw_conn->out.data_segment, |
513 | sg_list: sg, sg_count: count, offset, size: len, |
514 | NULL, hash: tx_hash); |
515 | } |
516 | |
517 | static void |
518 | iscsi_sw_tcp_send_linear_data_prep(struct iscsi_conn *conn, void *data, |
519 | size_t len) |
520 | { |
521 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
522 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
523 | struct ahash_request *tx_hash = NULL; |
524 | unsigned int hdr_spec_len; |
525 | |
526 | ISCSI_SW_TCP_DBG(conn, "datalen=%zd %s\n" , len, conn->datadgst_en ? |
527 | "digest enabled" : "digest disabled" ); |
528 | |
529 | /* Make sure the datalen matches what the caller |
530 | said he would send. */ |
531 | hdr_spec_len = ntoh24(tcp_sw_conn->out.hdr->dlength); |
532 | WARN_ON(iscsi_padded(len) != iscsi_padded(hdr_spec_len)); |
533 | |
534 | if (conn->datadgst_en) |
535 | tx_hash = tcp_sw_conn->tx_hash; |
536 | |
537 | iscsi_segment_init_linear(segment: &tcp_sw_conn->out.data_segment, |
538 | data, size: len, NULL, hash: tx_hash); |
539 | } |
540 | |
541 | static int iscsi_sw_tcp_pdu_init(struct iscsi_task *task, |
542 | unsigned int offset, unsigned int count) |
543 | { |
544 | struct iscsi_conn *conn = task->conn; |
545 | int err = 0; |
546 | |
547 | iscsi_sw_tcp_send_hdr_prep(conn, hdr: task->hdr, hdrlen: task->hdr_len); |
548 | |
549 | if (!count) |
550 | return 0; |
551 | |
552 | if (!task->sc) |
553 | iscsi_sw_tcp_send_linear_data_prep(conn, data: task->data, len: count); |
554 | else { |
555 | struct scsi_data_buffer *sdb = &task->sc->sdb; |
556 | |
557 | err = iscsi_sw_tcp_send_data_prep(conn, sg: sdb->table.sgl, |
558 | count: sdb->table.nents, offset, |
559 | len: count); |
560 | } |
561 | |
562 | if (err) { |
563 | /* got invalid offset/len */ |
564 | return -EIO; |
565 | } |
566 | return 0; |
567 | } |
568 | |
569 | static int iscsi_sw_tcp_pdu_alloc(struct iscsi_task *task, uint8_t opcode) |
570 | { |
571 | struct iscsi_tcp_task *tcp_task = task->dd_data; |
572 | |
573 | task->hdr = task->dd_data + sizeof(*tcp_task); |
574 | task->hdr_max = sizeof(struct iscsi_sw_tcp_hdrbuf) - ISCSI_DIGEST_SIZE; |
575 | return 0; |
576 | } |
577 | |
578 | static struct iscsi_cls_conn * |
579 | iscsi_sw_tcp_conn_create(struct iscsi_cls_session *cls_session, |
580 | uint32_t conn_idx) |
581 | { |
582 | struct iscsi_conn *conn; |
583 | struct iscsi_cls_conn *cls_conn; |
584 | struct iscsi_tcp_conn *tcp_conn; |
585 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
586 | struct crypto_ahash *tfm; |
587 | |
588 | cls_conn = iscsi_tcp_conn_setup(cls_session, dd_data_size: sizeof(*tcp_sw_conn), |
589 | conn_idx); |
590 | if (!cls_conn) |
591 | return NULL; |
592 | conn = cls_conn->dd_data; |
593 | tcp_conn = conn->dd_data; |
594 | tcp_sw_conn = tcp_conn->dd_data; |
595 | INIT_WORK(&conn->recvwork, iscsi_sw_tcp_recv_data_work); |
596 | tcp_sw_conn->queue_recv = iscsi_recv_from_iscsi_q; |
597 | |
598 | mutex_init(&tcp_sw_conn->sock_lock); |
599 | |
600 | tfm = crypto_alloc_ahash(alg_name: "crc32c" , type: 0, CRYPTO_ALG_ASYNC); |
601 | if (IS_ERR(ptr: tfm)) |
602 | goto free_conn; |
603 | |
604 | tcp_sw_conn->tx_hash = ahash_request_alloc(tfm, GFP_KERNEL); |
605 | if (!tcp_sw_conn->tx_hash) |
606 | goto free_tfm; |
607 | ahash_request_set_callback(req: tcp_sw_conn->tx_hash, flags: 0, NULL, NULL); |
608 | |
609 | tcp_sw_conn->rx_hash = ahash_request_alloc(tfm, GFP_KERNEL); |
610 | if (!tcp_sw_conn->rx_hash) |
611 | goto free_tx_hash; |
612 | ahash_request_set_callback(req: tcp_sw_conn->rx_hash, flags: 0, NULL, NULL); |
613 | |
614 | tcp_conn->rx_hash = tcp_sw_conn->rx_hash; |
615 | |
616 | return cls_conn; |
617 | |
618 | free_tx_hash: |
619 | ahash_request_free(req: tcp_sw_conn->tx_hash); |
620 | free_tfm: |
621 | crypto_free_ahash(tfm); |
622 | free_conn: |
623 | iscsi_conn_printk(KERN_ERR, conn, |
624 | "Could not create connection due to crc32c " |
625 | "loading error. Make sure the crc32c " |
626 | "module is built as a module or into the " |
627 | "kernel\n" ); |
628 | iscsi_tcp_conn_teardown(cls_conn); |
629 | return NULL; |
630 | } |
631 | |
632 | static void iscsi_sw_tcp_release_conn(struct iscsi_conn *conn) |
633 | { |
634 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
635 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
636 | struct socket *sock = tcp_sw_conn->sock; |
637 | |
638 | /* |
639 | * The iscsi transport class will make sure we are not called in |
640 | * parallel with start, stop, bind and destroys. However, this can be |
641 | * called twice if userspace does a stop then a destroy. |
642 | */ |
643 | if (!sock) |
644 | return; |
645 | |
646 | /* |
647 | * Make sure we start socket shutdown now in case userspace is up |
648 | * but delayed in releasing the socket. |
649 | */ |
650 | kernel_sock_shutdown(sock, how: SHUT_RDWR); |
651 | |
652 | sock_hold(sk: sock->sk); |
653 | iscsi_sw_tcp_conn_restore_callbacks(conn); |
654 | sock_put(sk: sock->sk); |
655 | |
656 | iscsi_suspend_rx(conn); |
657 | |
658 | mutex_lock(&tcp_sw_conn->sock_lock); |
659 | tcp_sw_conn->sock = NULL; |
660 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
661 | sockfd_put(sock); |
662 | } |
663 | |
664 | static void iscsi_sw_tcp_conn_destroy(struct iscsi_cls_conn *cls_conn) |
665 | { |
666 | struct iscsi_conn *conn = cls_conn->dd_data; |
667 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
668 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
669 | |
670 | iscsi_sw_tcp_release_conn(conn); |
671 | |
672 | ahash_request_free(req: tcp_sw_conn->rx_hash); |
673 | if (tcp_sw_conn->tx_hash) { |
674 | struct crypto_ahash *tfm; |
675 | |
676 | tfm = crypto_ahash_reqtfm(req: tcp_sw_conn->tx_hash); |
677 | ahash_request_free(req: tcp_sw_conn->tx_hash); |
678 | crypto_free_ahash(tfm); |
679 | } |
680 | |
681 | iscsi_tcp_conn_teardown(cls_conn); |
682 | } |
683 | |
684 | static void iscsi_sw_tcp_conn_stop(struct iscsi_cls_conn *cls_conn, int flag) |
685 | { |
686 | struct iscsi_conn *conn = cls_conn->dd_data; |
687 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
688 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
689 | struct socket *sock = tcp_sw_conn->sock; |
690 | |
691 | /* userspace may have goofed up and not bound us */ |
692 | if (!sock) |
693 | return; |
694 | |
695 | sock->sk->sk_err = EIO; |
696 | wake_up_interruptible(sk_sleep(sock->sk)); |
697 | |
698 | /* stop xmit side */ |
699 | iscsi_suspend_tx(conn); |
700 | |
701 | /* stop recv side and release socket */ |
702 | iscsi_sw_tcp_release_conn(conn); |
703 | |
704 | iscsi_conn_stop(cls_conn, flag); |
705 | } |
706 | |
707 | static int |
708 | iscsi_sw_tcp_conn_bind(struct iscsi_cls_session *cls_session, |
709 | struct iscsi_cls_conn *cls_conn, uint64_t transport_eph, |
710 | int is_leading) |
711 | { |
712 | struct iscsi_conn *conn = cls_conn->dd_data; |
713 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
714 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
715 | struct sock *sk; |
716 | struct socket *sock; |
717 | int err; |
718 | |
719 | /* lookup for existing socket */ |
720 | sock = sockfd_lookup(fd: (int)transport_eph, err: &err); |
721 | if (!sock) { |
722 | iscsi_conn_printk(KERN_ERR, conn, |
723 | "sockfd_lookup failed %d\n" , err); |
724 | return -EEXIST; |
725 | } |
726 | |
727 | err = -EINVAL; |
728 | if (!sk_is_tcp(sk: sock->sk)) |
729 | goto free_socket; |
730 | |
731 | err = iscsi_conn_bind(cls_session, cls_conn, is_leading); |
732 | if (err) |
733 | goto free_socket; |
734 | |
735 | mutex_lock(&tcp_sw_conn->sock_lock); |
736 | /* bind iSCSI connection and socket */ |
737 | tcp_sw_conn->sock = sock; |
738 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
739 | |
740 | /* setup Socket parameters */ |
741 | sk = sock->sk; |
742 | sk->sk_reuse = SK_CAN_REUSE; |
743 | sk->sk_sndtimeo = 15 * HZ; /* FIXME: make it configurable */ |
744 | sk->sk_allocation = GFP_ATOMIC; |
745 | sk->sk_use_task_frag = false; |
746 | sk_set_memalloc(sk); |
747 | sock_no_linger(sk); |
748 | |
749 | iscsi_sw_tcp_conn_set_callbacks(conn); |
750 | /* |
751 | * set receive state machine into initial state |
752 | */ |
753 | iscsi_tcp_hdr_recv_prep(tcp_conn); |
754 | return 0; |
755 | |
756 | free_socket: |
757 | sockfd_put(sock); |
758 | return err; |
759 | } |
760 | |
761 | static int iscsi_sw_tcp_conn_set_param(struct iscsi_cls_conn *cls_conn, |
762 | enum iscsi_param param, char *buf, |
763 | int buflen) |
764 | { |
765 | struct iscsi_conn *conn = cls_conn->dd_data; |
766 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
767 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
768 | |
769 | switch(param) { |
770 | case ISCSI_PARAM_HDRDGST_EN: |
771 | iscsi_set_param(cls_conn, param, buf, buflen); |
772 | break; |
773 | case ISCSI_PARAM_DATADGST_EN: |
774 | mutex_lock(&tcp_sw_conn->sock_lock); |
775 | if (!tcp_sw_conn->sock) { |
776 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
777 | return -ENOTCONN; |
778 | } |
779 | iscsi_set_param(cls_conn, param, buf, buflen); |
780 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
781 | break; |
782 | case ISCSI_PARAM_MAX_R2T: |
783 | return iscsi_tcp_set_max_r2t(conn, buf); |
784 | default: |
785 | return iscsi_set_param(cls_conn, param, buf, buflen); |
786 | } |
787 | |
788 | return 0; |
789 | } |
790 | |
791 | static int iscsi_sw_tcp_conn_get_param(struct iscsi_cls_conn *cls_conn, |
792 | enum iscsi_param param, char *buf) |
793 | { |
794 | struct iscsi_conn *conn = cls_conn->dd_data; |
795 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
796 | struct iscsi_tcp_conn *tcp_conn; |
797 | struct sockaddr_in6 addr; |
798 | struct socket *sock; |
799 | int rc; |
800 | |
801 | switch(param) { |
802 | case ISCSI_PARAM_CONN_PORT: |
803 | case ISCSI_PARAM_CONN_ADDRESS: |
804 | case ISCSI_PARAM_LOCAL_PORT: |
805 | spin_lock_bh(lock: &conn->session->frwd_lock); |
806 | if (!conn->session->leadconn) { |
807 | spin_unlock_bh(lock: &conn->session->frwd_lock); |
808 | return -ENOTCONN; |
809 | } |
810 | /* |
811 | * The conn has been setup and bound, so just grab a ref |
812 | * incase a destroy runs while we are in the net layer. |
813 | */ |
814 | iscsi_get_conn(conn: conn->cls_conn); |
815 | spin_unlock_bh(lock: &conn->session->frwd_lock); |
816 | |
817 | tcp_conn = conn->dd_data; |
818 | tcp_sw_conn = tcp_conn->dd_data; |
819 | |
820 | mutex_lock(&tcp_sw_conn->sock_lock); |
821 | sock = tcp_sw_conn->sock; |
822 | if (!sock) { |
823 | rc = -ENOTCONN; |
824 | goto sock_unlock; |
825 | } |
826 | |
827 | if (param == ISCSI_PARAM_LOCAL_PORT) |
828 | rc = kernel_getsockname(sock, |
829 | addr: (struct sockaddr *)&addr); |
830 | else |
831 | rc = kernel_getpeername(sock, |
832 | addr: (struct sockaddr *)&addr); |
833 | sock_unlock: |
834 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
835 | iscsi_put_conn(conn: conn->cls_conn); |
836 | if (rc < 0) |
837 | return rc; |
838 | |
839 | return iscsi_conn_get_addr_param(addr: (struct sockaddr_storage *) |
840 | &addr, param, buf); |
841 | default: |
842 | return iscsi_conn_get_param(cls_conn, param, buf); |
843 | } |
844 | |
845 | return 0; |
846 | } |
847 | |
848 | static int iscsi_sw_tcp_host_get_param(struct Scsi_Host *shost, |
849 | enum iscsi_host_param param, char *buf) |
850 | { |
851 | struct iscsi_sw_tcp_host *tcp_sw_host = iscsi_host_priv(shost); |
852 | struct iscsi_session *session; |
853 | struct iscsi_conn *conn; |
854 | struct iscsi_tcp_conn *tcp_conn; |
855 | struct iscsi_sw_tcp_conn *tcp_sw_conn; |
856 | struct sockaddr_in6 addr; |
857 | struct socket *sock; |
858 | int rc; |
859 | |
860 | switch (param) { |
861 | case ISCSI_HOST_PARAM_IPADDRESS: |
862 | session = tcp_sw_host->session; |
863 | if (!session) |
864 | return -ENOTCONN; |
865 | |
866 | spin_lock_bh(lock: &session->frwd_lock); |
867 | conn = session->leadconn; |
868 | if (!conn) { |
869 | spin_unlock_bh(lock: &session->frwd_lock); |
870 | return -ENOTCONN; |
871 | } |
872 | tcp_conn = conn->dd_data; |
873 | tcp_sw_conn = tcp_conn->dd_data; |
874 | /* |
875 | * The conn has been setup and bound, so just grab a ref |
876 | * incase a destroy runs while we are in the net layer. |
877 | */ |
878 | iscsi_get_conn(conn: conn->cls_conn); |
879 | spin_unlock_bh(lock: &session->frwd_lock); |
880 | |
881 | mutex_lock(&tcp_sw_conn->sock_lock); |
882 | sock = tcp_sw_conn->sock; |
883 | if (!sock) |
884 | rc = -ENOTCONN; |
885 | else |
886 | rc = kernel_getsockname(sock, addr: (struct sockaddr *)&addr); |
887 | mutex_unlock(lock: &tcp_sw_conn->sock_lock); |
888 | iscsi_put_conn(conn: conn->cls_conn); |
889 | if (rc < 0) |
890 | return rc; |
891 | |
892 | return iscsi_conn_get_addr_param(addr: (struct sockaddr_storage *) |
893 | &addr, |
894 | param: (enum iscsi_param)param, buf); |
895 | default: |
896 | return iscsi_host_get_param(shost, param, buf); |
897 | } |
898 | |
899 | return 0; |
900 | } |
901 | |
902 | static void |
903 | iscsi_sw_tcp_conn_get_stats(struct iscsi_cls_conn *cls_conn, |
904 | struct iscsi_stats *stats) |
905 | { |
906 | struct iscsi_conn *conn = cls_conn->dd_data; |
907 | struct iscsi_tcp_conn *tcp_conn = conn->dd_data; |
908 | struct iscsi_sw_tcp_conn *tcp_sw_conn = tcp_conn->dd_data; |
909 | |
910 | stats->custom_length = 3; |
911 | strcpy(p: stats->custom[0].desc, q: "tx_sendpage_failures" ); |
912 | stats->custom[0].value = tcp_sw_conn->sendpage_failures_cnt; |
913 | strcpy(p: stats->custom[1].desc, q: "rx_discontiguous_hdr" ); |
914 | stats->custom[1].value = tcp_sw_conn->discontiguous_hdr_cnt; |
915 | strcpy(p: stats->custom[2].desc, q: "eh_abort_cnt" ); |
916 | stats->custom[2].value = conn->eh_abort_cnt; |
917 | |
918 | iscsi_tcp_conn_get_stats(cls_conn, stats); |
919 | } |
920 | |
921 | static struct iscsi_cls_session * |
922 | iscsi_sw_tcp_session_create(struct iscsi_endpoint *ep, uint16_t cmds_max, |
923 | uint16_t qdepth, uint32_t initial_cmdsn) |
924 | { |
925 | struct iscsi_cls_session *cls_session; |
926 | struct iscsi_session *session; |
927 | struct iscsi_sw_tcp_host *tcp_sw_host; |
928 | struct Scsi_Host *shost; |
929 | int rc; |
930 | |
931 | if (ep) { |
932 | printk(KERN_ERR "iscsi_tcp: invalid ep %p.\n" , ep); |
933 | return NULL; |
934 | } |
935 | |
936 | shost = iscsi_host_alloc(sht: &iscsi_sw_tcp_sht, |
937 | dd_data_size: sizeof(struct iscsi_sw_tcp_host), xmit_can_sleep: 1); |
938 | if (!shost) |
939 | return NULL; |
940 | shost->transportt = iscsi_sw_tcp_scsi_transport; |
941 | shost->cmd_per_lun = qdepth; |
942 | shost->max_lun = iscsi_max_lun; |
943 | shost->max_id = 0; |
944 | shost->max_channel = 0; |
945 | shost->max_cmd_len = SCSI_MAX_VARLEN_CDB_SIZE; |
946 | |
947 | rc = iscsi_host_get_max_scsi_cmds(shost, requested_cmds_max: cmds_max); |
948 | if (rc < 0) |
949 | goto free_host; |
950 | shost->can_queue = rc; |
951 | |
952 | if (iscsi_host_add(shost, NULL)) |
953 | goto free_host; |
954 | |
955 | cls_session = iscsi_session_setup(&iscsi_sw_tcp_transport, shost, |
956 | cmds_max, 0, |
957 | sizeof(struct iscsi_tcp_task) + |
958 | sizeof(struct iscsi_sw_tcp_hdrbuf), |
959 | initial_cmdsn, 0); |
960 | if (!cls_session) |
961 | goto remove_host; |
962 | session = cls_session->dd_data; |
963 | |
964 | if (iscsi_tcp_r2tpool_alloc(session)) |
965 | goto remove_session; |
966 | |
967 | /* We are now fully setup so expose the session to sysfs. */ |
968 | tcp_sw_host = iscsi_host_priv(shost); |
969 | tcp_sw_host->session = session; |
970 | return cls_session; |
971 | |
972 | remove_session: |
973 | iscsi_session_teardown(cls_session); |
974 | remove_host: |
975 | iscsi_host_remove(shost, is_shutdown: false); |
976 | free_host: |
977 | iscsi_host_free(shost); |
978 | return NULL; |
979 | } |
980 | |
981 | static void iscsi_sw_tcp_session_destroy(struct iscsi_cls_session *cls_session) |
982 | { |
983 | struct Scsi_Host *shost = iscsi_session_to_shost(cls_session); |
984 | struct iscsi_session *session = cls_session->dd_data; |
985 | |
986 | if (WARN_ON_ONCE(session->leadconn)) |
987 | return; |
988 | |
989 | iscsi_session_remove(cls_session); |
990 | /* |
991 | * Our get_host_param needs to access the session, so remove the |
992 | * host from sysfs before freeing the session to make sure userspace |
993 | * is no longer accessing the callout. |
994 | */ |
995 | iscsi_host_remove(shost, is_shutdown: false); |
996 | |
997 | iscsi_tcp_r2tpool_free(session: cls_session->dd_data); |
998 | |
999 | iscsi_session_free(cls_session); |
1000 | iscsi_host_free(shost); |
1001 | } |
1002 | |
1003 | static umode_t iscsi_sw_tcp_attr_is_visible(int param_type, int param) |
1004 | { |
1005 | switch (param_type) { |
1006 | case ISCSI_HOST_PARAM: |
1007 | switch (param) { |
1008 | case ISCSI_HOST_PARAM_NETDEV_NAME: |
1009 | case ISCSI_HOST_PARAM_HWADDRESS: |
1010 | case ISCSI_HOST_PARAM_IPADDRESS: |
1011 | case ISCSI_HOST_PARAM_INITIATOR_NAME: |
1012 | return S_IRUGO; |
1013 | default: |
1014 | return 0; |
1015 | } |
1016 | case ISCSI_PARAM: |
1017 | switch (param) { |
1018 | case ISCSI_PARAM_MAX_RECV_DLENGTH: |
1019 | case ISCSI_PARAM_MAX_XMIT_DLENGTH: |
1020 | case ISCSI_PARAM_HDRDGST_EN: |
1021 | case ISCSI_PARAM_DATADGST_EN: |
1022 | case ISCSI_PARAM_CONN_ADDRESS: |
1023 | case ISCSI_PARAM_CONN_PORT: |
1024 | case ISCSI_PARAM_LOCAL_PORT: |
1025 | case ISCSI_PARAM_EXP_STATSN: |
1026 | case ISCSI_PARAM_PERSISTENT_ADDRESS: |
1027 | case ISCSI_PARAM_PERSISTENT_PORT: |
1028 | case ISCSI_PARAM_PING_TMO: |
1029 | case ISCSI_PARAM_RECV_TMO: |
1030 | case ISCSI_PARAM_INITIAL_R2T_EN: |
1031 | case ISCSI_PARAM_MAX_R2T: |
1032 | case ISCSI_PARAM_IMM_DATA_EN: |
1033 | case ISCSI_PARAM_FIRST_BURST: |
1034 | case ISCSI_PARAM_MAX_BURST: |
1035 | case ISCSI_PARAM_PDU_INORDER_EN: |
1036 | case ISCSI_PARAM_DATASEQ_INORDER_EN: |
1037 | case ISCSI_PARAM_ERL: |
1038 | case ISCSI_PARAM_TARGET_NAME: |
1039 | case ISCSI_PARAM_TPGT: |
1040 | case ISCSI_PARAM_USERNAME: |
1041 | case ISCSI_PARAM_PASSWORD: |
1042 | case ISCSI_PARAM_USERNAME_IN: |
1043 | case ISCSI_PARAM_PASSWORD_IN: |
1044 | case ISCSI_PARAM_FAST_ABORT: |
1045 | case ISCSI_PARAM_ABORT_TMO: |
1046 | case ISCSI_PARAM_LU_RESET_TMO: |
1047 | case ISCSI_PARAM_TGT_RESET_TMO: |
1048 | case ISCSI_PARAM_IFACE_NAME: |
1049 | case ISCSI_PARAM_INITIATOR_NAME: |
1050 | return S_IRUGO; |
1051 | default: |
1052 | return 0; |
1053 | } |
1054 | } |
1055 | |
1056 | return 0; |
1057 | } |
1058 | |
1059 | static int iscsi_sw_tcp_slave_configure(struct scsi_device *sdev) |
1060 | { |
1061 | struct iscsi_sw_tcp_host *tcp_sw_host = iscsi_host_priv(sdev->host); |
1062 | struct iscsi_session *session = tcp_sw_host->session; |
1063 | struct iscsi_conn *conn = session->leadconn; |
1064 | |
1065 | if (conn->datadgst_en) |
1066 | blk_queue_flag_set(QUEUE_FLAG_STABLE_WRITES, |
1067 | q: sdev->request_queue); |
1068 | blk_queue_dma_alignment(sdev->request_queue, 0); |
1069 | return 0; |
1070 | } |
1071 | |
1072 | static const struct scsi_host_template iscsi_sw_tcp_sht = { |
1073 | .module = THIS_MODULE, |
1074 | .name = "iSCSI Initiator over TCP/IP" , |
1075 | .queuecommand = iscsi_queuecommand, |
1076 | .change_queue_depth = scsi_change_queue_depth, |
1077 | .can_queue = ISCSI_TOTAL_CMDS_MAX, |
1078 | .sg_tablesize = 4096, |
1079 | .max_sectors = 0xFFFF, |
1080 | .cmd_per_lun = ISCSI_DEF_CMD_PER_LUN, |
1081 | .eh_timed_out = iscsi_eh_cmd_timed_out, |
1082 | .eh_abort_handler = iscsi_eh_abort, |
1083 | .eh_device_reset_handler= iscsi_eh_device_reset, |
1084 | .eh_target_reset_handler = iscsi_eh_recover_target, |
1085 | .dma_boundary = PAGE_SIZE - 1, |
1086 | .slave_configure = iscsi_sw_tcp_slave_configure, |
1087 | .proc_name = "iscsi_tcp" , |
1088 | .this_id = -1, |
1089 | .track_queue_depth = 1, |
1090 | .cmd_size = sizeof(struct iscsi_cmd), |
1091 | }; |
1092 | |
1093 | static struct iscsi_transport iscsi_sw_tcp_transport = { |
1094 | .owner = THIS_MODULE, |
1095 | .name = "tcp" , |
1096 | .caps = CAP_RECOVERY_L0 | CAP_MULTI_R2T | CAP_HDRDGST |
1097 | | CAP_DATADGST, |
1098 | /* session management */ |
1099 | .create_session = iscsi_sw_tcp_session_create, |
1100 | .destroy_session = iscsi_sw_tcp_session_destroy, |
1101 | /* connection management */ |
1102 | .create_conn = iscsi_sw_tcp_conn_create, |
1103 | .bind_conn = iscsi_sw_tcp_conn_bind, |
1104 | .destroy_conn = iscsi_sw_tcp_conn_destroy, |
1105 | .attr_is_visible = iscsi_sw_tcp_attr_is_visible, |
1106 | .set_param = iscsi_sw_tcp_conn_set_param, |
1107 | .get_conn_param = iscsi_sw_tcp_conn_get_param, |
1108 | .get_session_param = iscsi_session_get_param, |
1109 | .start_conn = iscsi_conn_start, |
1110 | .stop_conn = iscsi_sw_tcp_conn_stop, |
1111 | /* iscsi host params */ |
1112 | .get_host_param = iscsi_sw_tcp_host_get_param, |
1113 | .set_host_param = iscsi_host_set_param, |
1114 | /* IO */ |
1115 | .send_pdu = iscsi_conn_send_pdu, |
1116 | .get_stats = iscsi_sw_tcp_conn_get_stats, |
1117 | /* iscsi task/cmd helpers */ |
1118 | .init_task = iscsi_tcp_task_init, |
1119 | .xmit_task = iscsi_tcp_task_xmit, |
1120 | .cleanup_task = iscsi_tcp_cleanup_task, |
1121 | /* low level pdu helpers */ |
1122 | .xmit_pdu = iscsi_sw_tcp_pdu_xmit, |
1123 | .init_pdu = iscsi_sw_tcp_pdu_init, |
1124 | .alloc_pdu = iscsi_sw_tcp_pdu_alloc, |
1125 | /* recovery */ |
1126 | .session_recovery_timedout = iscsi_session_recovery_timedout, |
1127 | }; |
1128 | |
1129 | static int __init iscsi_sw_tcp_init(void) |
1130 | { |
1131 | if (iscsi_max_lun < 1) { |
1132 | printk(KERN_ERR "iscsi_tcp: Invalid max_lun value of %u\n" , |
1133 | iscsi_max_lun); |
1134 | return -EINVAL; |
1135 | } |
1136 | |
1137 | iscsi_sw_tcp_scsi_transport = iscsi_register_transport( |
1138 | tt: &iscsi_sw_tcp_transport); |
1139 | if (!iscsi_sw_tcp_scsi_transport) |
1140 | return -ENODEV; |
1141 | |
1142 | return 0; |
1143 | } |
1144 | |
1145 | static void __exit iscsi_sw_tcp_exit(void) |
1146 | { |
1147 | iscsi_unregister_transport(tt: &iscsi_sw_tcp_transport); |
1148 | } |
1149 | |
1150 | module_init(iscsi_sw_tcp_init); |
1151 | module_exit(iscsi_sw_tcp_exit); |
1152 | |