]> bbs.cooldavid.org Git - net-next-2.6.git/blame - drivers/scsi/libfc/fc_lport.c
Merge branches 'sh/pio-death', 'sh/nommu', 'sh/clkfwk', 'sh/core' and 'sh/intc-extens...
[net-next-2.6.git] / drivers / scsi / libfc / fc_lport.c
CommitLineData
42e9a92f
RL
1/*
2 * Copyright(c) 2007 Intel Corporation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
11 * more details.
12 *
13 * You should have received a copy of the GNU General Public License along with
14 * this program; if not, write to the Free Software Foundation, Inc.,
15 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
16 *
17 * Maintained at www.Open-FCoE.org
18 */
19
20/*
21 * PORT LOCKING NOTES
22 *
23 * These comments only apply to the 'port code' which consists of the lport,
24 * disc and rport blocks.
25 *
26 * MOTIVATION
27 *
28 * The lport, disc and rport blocks all have mutexes that are used to protect
29 * those objects. The main motivation for these locks is to prevent from
30 * having an lport reset just before we send a frame. In that scenario the
31 * lport's FID would get set to zero and then we'd send a frame with an
32 * invalid SID. We also need to ensure that states don't change unexpectedly
33 * while processing another state.
34 *
732bee7a 35 * HIERARCHY
42e9a92f 36 *
732bee7a 37 * The following hierarchy defines the locking rules. A greater lock
42e9a92f 38 * may be held before acquiring a lesser lock, but a lesser lock should never
732bee7a 39 * be held while attempting to acquire a greater lock. Here is the hierarchy-
42e9a92f
RL
40 *
41 * lport > disc, lport > rport, disc > rport
42 *
43 * CALLBACKS
44 *
45 * The callbacks cause complications with this scheme. There is a callback
46 * from the rport (to either lport or disc) and a callback from disc
47 * (to the lport).
48 *
49 * As rports exit the rport state machine a callback is made to the owner of
50 * the rport to notify success or failure. Since the callback is likely to
51 * cause the lport or disc to grab its lock we cannot hold the rport lock
52 * while making the callback. To ensure that the rport is not free'd while
53 * processing the callback the rport callbacks are serialized through a
54 * single-threaded workqueue. An rport would never be free'd while in a
55 * callback handler becuase no other rport work in this queue can be executed
56 * at the same time.
57 *
58 * When discovery succeeds or fails a callback is made to the lport as
af901ca1 59 * notification. Currently, successful discovery causes the lport to take no
42e9a92f
RL
60 * action. A failure will cause the lport to reset. There is likely a circular
61 * locking problem with this implementation.
62 */
63
64/*
65 * LPORT LOCKING
66 *
67 * The critical sections protected by the lport's mutex are quite broad and
68 * may be improved upon in the future. The lport code and its locking doesn't
69 * influence the I/O path, so excessive locking doesn't penalize I/O
70 * performance.
71 *
72 * The strategy is to lock whenever processing a request or response. Note
73 * that every _enter_* function corresponds to a state change. They generally
74 * change the lports state and then send a request out on the wire. We lock
75 * before calling any of these functions to protect that state change. This
76 * means that the entry points into the lport block manage the locks while
77 * the state machine can transition between states (i.e. _enter_* functions)
78 * while always staying protected.
79 *
80 * When handling responses we also hold the lport mutex broadly. When the
81 * lport receives the response frame it locks the mutex and then calls the
82 * appropriate handler for the particuar response. Generally a response will
83 * trigger a state change and so the lock must already be held.
84 *
85 * Retries also have to consider the locking. The retries occur from a work
86 * context and the work function will lock the lport and then retry the state
87 * (i.e. _enter_* function).
88 */
89
90#include <linux/timer.h>
5a0e3ad6 91#include <linux/slab.h>
42e9a92f
RL
92#include <asm/unaligned.h>
93
94#include <scsi/fc/fc_gs.h>
95
96#include <scsi/libfc.h>
97#include <scsi/fc_encode.h>
a51ab396 98#include <linux/scatterlist.h>
42e9a92f 99
8866a5d9
RL
100#include "fc_libfc.h"
101
42e9a92f
RL
102/* Fabric IDs to use for point-to-point mode, chosen on whims. */
103#define FC_LOCAL_PTP_FID_LO 0x010101
104#define FC_LOCAL_PTP_FID_HI 0x010102
105
106#define DNS_DELAY 3 /* Discovery delay after RSCN (in seconds)*/
107
42e9a92f
RL
108static void fc_lport_error(struct fc_lport *, struct fc_frame *);
109
110static void fc_lport_enter_reset(struct fc_lport *);
111static void fc_lport_enter_flogi(struct fc_lport *);
112static void fc_lport_enter_dns(struct fc_lport *);
c914f7d1 113static void fc_lport_enter_ns(struct fc_lport *, enum fc_lport_state);
42e9a92f
RL
114static void fc_lport_enter_scr(struct fc_lport *);
115static void fc_lport_enter_ready(struct fc_lport *);
116static void fc_lport_enter_logo(struct fc_lport *);
117
118static const char *fc_lport_state_names[] = {
b1d9fd55 119 [LPORT_ST_DISABLED] = "disabled",
42e9a92f
RL
120 [LPORT_ST_FLOGI] = "FLOGI",
121 [LPORT_ST_DNS] = "dNS",
c9c7bd7a 122 [LPORT_ST_RNN_ID] = "RNN_ID",
5baa17c3 123 [LPORT_ST_RSNN_NN] = "RSNN_NN",
c9866a54 124 [LPORT_ST_RSPN_ID] = "RSPN_ID",
42e9a92f 125 [LPORT_ST_RFT_ID] = "RFT_ID",
ab593b18 126 [LPORT_ST_RFF_ID] = "RFF_ID",
42e9a92f
RL
127 [LPORT_ST_SCR] = "SCR",
128 [LPORT_ST_READY] = "Ready",
129 [LPORT_ST_LOGO] = "LOGO",
130 [LPORT_ST_RESET] = "reset",
131};
132
a51ab396
SM
133/**
134 * struct fc_bsg_info - FC Passthrough managemet structure
135 * @job: The passthrough job
136 * @lport: The local port to pass through a command
137 * @rsp_code: The expected response code
3a3b42bf 138 * @sg: job->reply_payload.sg_list
a51ab396
SM
139 * @nents: job->reply_payload.sg_cnt
140 * @offset: The offset into the response data
141 */
142struct fc_bsg_info {
143 struct fc_bsg_job *job;
144 struct fc_lport *lport;
145 u16 rsp_code;
146 struct scatterlist *sg;
147 u32 nents;
148 size_t offset;
149};
150
3a3b42bf
RL
151/**
152 * fc_frame_drop() - Dummy frame handler
153 * @lport: The local port the frame was received on
154 * @fp: The received frame
155 */
42e9a92f
RL
156static int fc_frame_drop(struct fc_lport *lport, struct fc_frame *fp)
157{
158 fc_frame_free(fp);
159 return 0;
160}
161
162/**
34f42a07 163 * fc_lport_rport_callback() - Event handler for rport events
42e9a92f 164 * @lport: The lport which is receiving the event
9fb9d328 165 * @rdata: private remote port data
42e9a92f
RL
166 * @event: The event that occured
167 *
168 * Locking Note: The rport lock should not be held when calling
169 * this function.
170 */
171static void fc_lport_rport_callback(struct fc_lport *lport,
9fb9d328 172 struct fc_rport_priv *rdata,
42e9a92f
RL
173 enum fc_rport_event event)
174{
ce8b5df0 175 FC_LPORT_DBG(lport, "Received a %d event for port (%6.6x)\n", event,
f211fa51 176 rdata->ids.port_id);
42e9a92f 177
b5cbf083 178 mutex_lock(&lport->lp_mutex);
42e9a92f 179 switch (event) {
4c0f62b5 180 case RPORT_EV_READY:
b5cbf083 181 if (lport->state == LPORT_ST_DNS) {
3a3b42bf 182 lport->dns_rdata = rdata;
c914f7d1 183 fc_lport_enter_ns(lport, LPORT_ST_RNN_ID);
b5cbf083
JE
184 } else {
185 FC_LPORT_DBG(lport, "Received an READY event "
ce8b5df0 186 "on port (%6.6x) for the directory "
b5cbf083
JE
187 "server, but the lport is not "
188 "in the DNS state, it's in the "
189 "%d state", rdata->ids.port_id,
190 lport->state);
191 lport->tt.rport_logoff(rdata);
192 }
42e9a92f
RL
193 break;
194 case RPORT_EV_LOGO:
195 case RPORT_EV_FAILED:
196 case RPORT_EV_STOP:
3a3b42bf 197 lport->dns_rdata = NULL;
42e9a92f
RL
198 break;
199 case RPORT_EV_NONE:
200 break;
201 }
b5cbf083 202 mutex_unlock(&lport->lp_mutex);
42e9a92f
RL
203}
204
205/**
34f42a07 206 * fc_lport_state() - Return a string which represents the lport's state
42e9a92f
RL
207 * @lport: The lport whose state is to converted to a string
208 */
209static const char *fc_lport_state(struct fc_lport *lport)
210{
211 const char *cp;
212
213 cp = fc_lport_state_names[lport->state];
214 if (!cp)
215 cp = "unknown";
216 return cp;
217}
218
219/**
34f42a07 220 * fc_lport_ptp_setup() - Create an rport for point-to-point mode
3a3b42bf
RL
221 * @lport: The lport to attach the ptp rport to
222 * @remote_fid: The FID of the ptp rport
42e9a92f
RL
223 * @remote_wwpn: The WWPN of the ptp rport
224 * @remote_wwnn: The WWNN of the ptp rport
225 */
226static void fc_lport_ptp_setup(struct fc_lport *lport,
227 u32 remote_fid, u64 remote_wwpn,
228 u64 remote_wwnn)
229{
48f00902 230 mutex_lock(&lport->disc.disc_mutex);
2f2ac4a0 231 if (lport->ptp_rdata) {
3a3b42bf 232 lport->tt.rport_logoff(lport->ptp_rdata);
2f2ac4a0
JE
233 kref_put(&lport->ptp_rdata->kref, lport->tt.rport_destroy);
234 }
3a3b42bf 235 lport->ptp_rdata = lport->tt.rport_create(lport, remote_fid);
2f2ac4a0 236 kref_get(&lport->ptp_rdata->kref);
3a3b42bf
RL
237 lport->ptp_rdata->ids.port_name = remote_wwpn;
238 lport->ptp_rdata->ids.node_name = remote_wwnn;
48f00902 239 mutex_unlock(&lport->disc.disc_mutex);
42e9a92f 240
3a3b42bf 241 lport->tt.rport_login(lport->ptp_rdata);
42e9a92f
RL
242
243 fc_lport_enter_ready(lport);
244}
245
3a3b42bf
RL
246/**
247 * fc_get_host_port_state() - Return the port state of the given Scsi_Host
248 * @shost: The SCSI host whose port state is to be determined
249 */
42e9a92f
RL
250void fc_get_host_port_state(struct Scsi_Host *shost)
251{
3a3b42bf 252 struct fc_lport *lport = shost_priv(shost);
42e9a92f 253
3a3b42bf
RL
254 mutex_lock(&lport->lp_mutex);
255 if (!lport->link_up)
8faecddb 256 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
42e9a92f 257 else
3a3b42bf 258 switch (lport->state) {
8faecddb
CL
259 case LPORT_ST_READY:
260 fc_host_port_state(shost) = FC_PORTSTATE_ONLINE;
261 break;
262 default:
263 fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE;
264 }
3a3b42bf 265 mutex_unlock(&lport->lp_mutex);
42e9a92f
RL
266}
267EXPORT_SYMBOL(fc_get_host_port_state);
268
3a3b42bf
RL
269/**
270 * fc_get_host_speed() - Return the speed of the given Scsi_Host
271 * @shost: The SCSI host whose port speed is to be determined
272 */
42e9a92f
RL
273void fc_get_host_speed(struct Scsi_Host *shost)
274{
275 struct fc_lport *lport = shost_priv(shost);
276
277 fc_host_speed(shost) = lport->link_speed;
278}
279EXPORT_SYMBOL(fc_get_host_speed);
280
3a3b42bf
RL
281/**
282 * fc_get_host_stats() - Return the Scsi_Host's statistics
283 * @shost: The SCSI host whose statistics are to be returned
284 */
42e9a92f
RL
285struct fc_host_statistics *fc_get_host_stats(struct Scsi_Host *shost)
286{
42e9a92f 287 struct fc_host_statistics *fcoe_stats;
3a3b42bf 288 struct fc_lport *lport = shost_priv(shost);
42e9a92f 289 struct timespec v0, v1;
582b45bc 290 unsigned int cpu;
42e9a92f 291
3a3b42bf 292 fcoe_stats = &lport->host_stats;
42e9a92f
RL
293 memset(fcoe_stats, 0, sizeof(struct fc_host_statistics));
294
295 jiffies_to_timespec(jiffies, &v0);
3a3b42bf 296 jiffies_to_timespec(lport->boot_time, &v1);
42e9a92f
RL
297 fcoe_stats->seconds_since_last_reset = (v0.tv_sec - v1.tv_sec);
298
582b45bc
RL
299 for_each_possible_cpu(cpu) {
300 struct fcoe_dev_stats *stats;
301
3a3b42bf 302 stats = per_cpu_ptr(lport->dev_stats, cpu);
582b45bc 303
42e9a92f
RL
304 fcoe_stats->tx_frames += stats->TxFrames;
305 fcoe_stats->tx_words += stats->TxWords;
306 fcoe_stats->rx_frames += stats->RxFrames;
307 fcoe_stats->rx_words += stats->RxWords;
308 fcoe_stats->error_frames += stats->ErrorFrames;
309 fcoe_stats->invalid_crc_count += stats->InvalidCRCCount;
310 fcoe_stats->fcp_input_requests += stats->InputRequests;
311 fcoe_stats->fcp_output_requests += stats->OutputRequests;
312 fcoe_stats->fcp_control_requests += stats->ControlRequests;
313 fcoe_stats->fcp_input_megabytes += stats->InputMegabytes;
314 fcoe_stats->fcp_output_megabytes += stats->OutputMegabytes;
315 fcoe_stats->link_failure_count += stats->LinkFailureCount;
316 }
317 fcoe_stats->lip_count = -1;
318 fcoe_stats->nos_count = -1;
319 fcoe_stats->loss_of_sync_count = -1;
320 fcoe_stats->loss_of_signal_count = -1;
321 fcoe_stats->prim_seq_protocol_err_count = -1;
322 fcoe_stats->dumped_frames = -1;
323 return fcoe_stats;
324}
325EXPORT_SYMBOL(fc_get_host_stats);
326
3a3b42bf
RL
327/**
328 * fc_lport_flogi_fill() - Fill in FLOGI command for request
329 * @lport: The local port the FLOGI is for
330 * @flogi: The FLOGI command
331 * @op: The opcode
42e9a92f 332 */
3a3b42bf
RL
333static void fc_lport_flogi_fill(struct fc_lport *lport,
334 struct fc_els_flogi *flogi,
335 unsigned int op)
42e9a92f
RL
336{
337 struct fc_els_csp *sp;
338 struct fc_els_cssp *cp;
339
340 memset(flogi, 0, sizeof(*flogi));
341 flogi->fl_cmd = (u8) op;
342 put_unaligned_be64(lport->wwpn, &flogi->fl_wwpn);
343 put_unaligned_be64(lport->wwnn, &flogi->fl_wwnn);
344 sp = &flogi->fl_csp;
345 sp->sp_hi_ver = 0x20;
346 sp->sp_lo_ver = 0x20;
347 sp->sp_bb_cred = htons(10); /* this gets set by gateway */
348 sp->sp_bb_data = htons((u16) lport->mfs);
349 cp = &flogi->fl_cssp[3 - 1]; /* class 3 parameters */
350 cp->cp_class = htons(FC_CPC_VALID | FC_CPC_SEQ);
351 if (op != ELS_FLOGI) {
352 sp->sp_features = htons(FC_SP_FT_CIRO);
353 sp->sp_tot_seq = htons(255); /* seq. we accept */
354 sp->sp_rel_off = htons(0x1f);
355 sp->sp_e_d_tov = htonl(lport->e_d_tov);
356
357 cp->cp_rdfs = htons((u16) lport->mfs);
358 cp->cp_con_seq = htons(255);
359 cp->cp_open_seq = 1;
360 }
361}
362
3a3b42bf
RL
363/**
364 * fc_lport_add_fc4_type() - Add a supported FC-4 type to a local port
365 * @lport: The local port to add a new FC-4 type to
366 * @type: The new FC-4 type
42e9a92f
RL
367 */
368static void fc_lport_add_fc4_type(struct fc_lport *lport, enum fc_fh_type type)
369{
370 __be32 *mp;
371
372 mp = &lport->fcts.ff_type_map[type / FC_NS_BPW];
373 *mp = htonl(ntohl(*mp) | 1UL << (type % FC_NS_BPW));
374}
375
376/**
34f42a07 377 * fc_lport_recv_rlir_req() - Handle received Registered Link Incident Report.
42e9a92f 378 * @lport: Fibre Channel local port recieving the RLIR
92261156 379 * @fp: The RLIR request frame
42e9a92f 380 *
1b69bc06 381 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
382 * this function.
383 */
92261156 384static void fc_lport_recv_rlir_req(struct fc_lport *lport, struct fc_frame *fp)
42e9a92f 385{
7414705e
RL
386 FC_LPORT_DBG(lport, "Received RLIR request while in state %s\n",
387 fc_lport_state(lport));
42e9a92f 388
92261156 389 lport->tt.seq_els_rsp_send(fp, ELS_LS_ACC, NULL);
42e9a92f
RL
390 fc_frame_free(fp);
391}
392
393/**
34f42a07 394 * fc_lport_recv_echo_req() - Handle received ECHO request
3a3b42bf 395 * @lport: The local port recieving the ECHO
92261156 396 * @fp: ECHO request frame
42e9a92f 397 *
1b69bc06 398 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
399 * this function.
400 */
92261156
JE
401static void fc_lport_recv_echo_req(struct fc_lport *lport,
402 struct fc_frame *in_fp)
42e9a92f
RL
403{
404 struct fc_frame *fp;
42e9a92f
RL
405 unsigned int len;
406 void *pp;
407 void *dp;
42e9a92f 408
1b69bc06 409 FC_LPORT_DBG(lport, "Received ECHO request while in state %s\n",
7414705e 410 fc_lport_state(lport));
42e9a92f
RL
411
412 len = fr_len(in_fp) - sizeof(struct fc_frame_header);
413 pp = fc_frame_payload_get(in_fp, len);
414
415 if (len < sizeof(__be32))
416 len = sizeof(__be32);
417
418 fp = fc_frame_alloc(lport, len);
419 if (fp) {
420 dp = fc_frame_payload_get(fp, len);
421 memcpy(dp, pp, len);
1b69bc06 422 *((__be32 *)dp) = htonl(ELS_LS_ACC << 24);
24f089e2
JE
423 fc_fill_reply_hdr(fp, in_fp, FC_RCTL_ELS_REP, 0);
424 lport->tt.frame_send(lport, fp);
42e9a92f
RL
425 }
426 fc_frame_free(in_fp);
427}
428
429/**
1b69bc06 430 * fc_lport_recv_rnid_req() - Handle received Request Node ID data request
1b69bc06 431 * @lport: The local port recieving the RNID
92261156 432 * @fp: The RNID request frame
42e9a92f 433 *
1b69bc06 434 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
435 * this function.
436 */
92261156
JE
437static void fc_lport_recv_rnid_req(struct fc_lport *lport,
438 struct fc_frame *in_fp)
42e9a92f
RL
439{
440 struct fc_frame *fp;
42e9a92f
RL
441 struct fc_els_rnid *req;
442 struct {
443 struct fc_els_rnid_resp rnid;
444 struct fc_els_rnid_cid cid;
445 struct fc_els_rnid_gen gen;
446 } *rp;
447 struct fc_seq_els_data rjt_data;
448 u8 fmt;
449 size_t len;
42e9a92f 450
7414705e
RL
451 FC_LPORT_DBG(lport, "Received RNID request while in state %s\n",
452 fc_lport_state(lport));
42e9a92f
RL
453
454 req = fc_frame_payload_get(in_fp, sizeof(*req));
455 if (!req) {
42e9a92f
RL
456 rjt_data.reason = ELS_RJT_LOGIC;
457 rjt_data.explan = ELS_EXPL_NONE;
92261156 458 lport->tt.seq_els_rsp_send(in_fp, ELS_LS_RJT, &rjt_data);
42e9a92f
RL
459 } else {
460 fmt = req->rnid_fmt;
461 len = sizeof(*rp);
462 if (fmt != ELS_RNIDF_GEN ||
463 ntohl(lport->rnid_gen.rnid_atype) == 0) {
464 fmt = ELS_RNIDF_NONE; /* nothing to provide */
465 len -= sizeof(rp->gen);
466 }
467 fp = fc_frame_alloc(lport, len);
468 if (fp) {
469 rp = fc_frame_payload_get(fp, len);
470 memset(rp, 0, len);
471 rp->rnid.rnid_cmd = ELS_LS_ACC;
472 rp->rnid.rnid_fmt = fmt;
473 rp->rnid.rnid_cid_len = sizeof(rp->cid);
474 rp->cid.rnid_wwpn = htonll(lport->wwpn);
475 rp->cid.rnid_wwnn = htonll(lport->wwnn);
476 if (fmt == ELS_RNIDF_GEN) {
477 rp->rnid.rnid_sid_len = sizeof(rp->gen);
478 memcpy(&rp->gen, &lport->rnid_gen,
479 sizeof(rp->gen));
480 }
24f089e2
JE
481 fc_fill_reply_hdr(fp, in_fp, FC_RCTL_ELS_REP, 0);
482 lport->tt.frame_send(lport, fp);
42e9a92f
RL
483 }
484 }
485 fc_frame_free(in_fp);
486}
487
42e9a92f 488/**
34f42a07 489 * fc_lport_recv_logo_req() - Handle received fabric LOGO request
3a3b42bf 490 * @lport: The local port recieving the LOGO
92261156 491 * @fp: The LOGO request frame
42e9a92f
RL
492 *
493 * Locking Note: The lport lock is exected to be held before calling
494 * this function.
495 */
92261156 496static void fc_lport_recv_logo_req(struct fc_lport *lport, struct fc_frame *fp)
42e9a92f 497{
92261156 498 lport->tt.seq_els_rsp_send(fp, ELS_LS_ACC, NULL);
42e9a92f
RL
499 fc_lport_enter_reset(lport);
500 fc_frame_free(fp);
501}
502
503/**
34f42a07 504 * fc_fabric_login() - Start the lport state machine
3a3b42bf 505 * @lport: The local port that should log into the fabric
42e9a92f
RL
506 *
507 * Locking Note: This function should not be called
508 * with the lport lock held.
509 */
510int fc_fabric_login(struct fc_lport *lport)
511{
512 int rc = -1;
513
514 mutex_lock(&lport->lp_mutex);
55a66d3c
VD
515 if (lport->state == LPORT_ST_DISABLED ||
516 lport->state == LPORT_ST_LOGO) {
517 fc_lport_state_enter(lport, LPORT_ST_RESET);
42e9a92f
RL
518 fc_lport_enter_reset(lport);
519 rc = 0;
520 }
521 mutex_unlock(&lport->lp_mutex);
522
523 return rc;
524}
525EXPORT_SYMBOL(fc_fabric_login);
526
527/**
8faecddb 528 * __fc_linkup() - Handler for transport linkup events
42e9a92f 529 * @lport: The lport whose link is up
8faecddb
CL
530 *
531 * Locking: must be called with the lp_mutex held
42e9a92f 532 */
8faecddb 533void __fc_linkup(struct fc_lport *lport)
42e9a92f 534{
bc0e17f6
VD
535 if (!lport->link_up) {
536 lport->link_up = 1;
42e9a92f
RL
537
538 if (lport->state == LPORT_ST_RESET)
539 fc_lport_enter_flogi(lport);
540 }
8faecddb
CL
541}
542
543/**
544 * fc_linkup() - Handler for transport linkup events
3a3b42bf 545 * @lport: The local port whose link is up
8faecddb
CL
546 */
547void fc_linkup(struct fc_lport *lport)
548{
ce8b5df0 549 printk(KERN_INFO "host%d: libfc: Link up on port (%6.6x)\n",
7b2787ec 550 lport->host->host_no, lport->port_id);
8faecddb
CL
551
552 mutex_lock(&lport->lp_mutex);
553 __fc_linkup(lport);
42e9a92f
RL
554 mutex_unlock(&lport->lp_mutex);
555}
556EXPORT_SYMBOL(fc_linkup);
557
558/**
8faecddb 559 * __fc_linkdown() - Handler for transport linkdown events
42e9a92f 560 * @lport: The lport whose link is down
8faecddb
CL
561 *
562 * Locking: must be called with the lp_mutex held
42e9a92f 563 */
8faecddb 564void __fc_linkdown(struct fc_lport *lport)
42e9a92f 565{
bc0e17f6
VD
566 if (lport->link_up) {
567 lport->link_up = 0;
42e9a92f
RL
568 fc_lport_enter_reset(lport);
569 lport->tt.fcp_cleanup(lport);
570 }
8faecddb
CL
571}
572
573/**
574 * fc_linkdown() - Handler for transport linkdown events
3a3b42bf 575 * @lport: The local port whose link is down
8faecddb
CL
576 */
577void fc_linkdown(struct fc_lport *lport)
578{
ce8b5df0 579 printk(KERN_INFO "host%d: libfc: Link down on port (%6.6x)\n",
7b2787ec 580 lport->host->host_no, lport->port_id);
8faecddb
CL
581
582 mutex_lock(&lport->lp_mutex);
583 __fc_linkdown(lport);
42e9a92f
RL
584 mutex_unlock(&lport->lp_mutex);
585}
586EXPORT_SYMBOL(fc_linkdown);
587
42e9a92f 588/**
34f42a07 589 * fc_fabric_logoff() - Logout of the fabric
3a3b42bf 590 * @lport: The local port to logoff the fabric
42e9a92f
RL
591 *
592 * Return value:
593 * 0 for success, -1 for failure
34f42a07 594 */
42e9a92f
RL
595int fc_fabric_logoff(struct fc_lport *lport)
596{
597 lport->tt.disc_stop_final(lport);
598 mutex_lock(&lport->lp_mutex);
3a3b42bf
RL
599 if (lport->dns_rdata)
600 lport->tt.rport_logoff(lport->dns_rdata);
a0fd2e49
AJ
601 mutex_unlock(&lport->lp_mutex);
602 lport->tt.rport_flush_queue();
603 mutex_lock(&lport->lp_mutex);
42e9a92f
RL
604 fc_lport_enter_logo(lport);
605 mutex_unlock(&lport->lp_mutex);
f7db2c15 606 cancel_delayed_work_sync(&lport->retry_work);
42e9a92f
RL
607 return 0;
608}
609EXPORT_SYMBOL(fc_fabric_logoff);
610
611/**
3a3b42bf
RL
612 * fc_lport_destroy() - Unregister a fc_lport
613 * @lport: The local port to unregister
42e9a92f 614 *
42e9a92f
RL
615 * Note:
616 * exit routine for fc_lport instance
617 * clean-up all the allocated memory
618 * and free up other system resources.
619 *
34f42a07 620 */
42e9a92f
RL
621int fc_lport_destroy(struct fc_lport *lport)
622{
bbf15669 623 mutex_lock(&lport->lp_mutex);
b1d9fd55 624 lport->state = LPORT_ST_DISABLED;
bbf15669 625 lport->link_up = 0;
42e9a92f 626 lport->tt.frame_send = fc_frame_drop;
bbf15669
AJ
627 mutex_unlock(&lport->lp_mutex);
628
42e9a92f 629 lport->tt.fcp_abort_io(lport);
e9ba8b42 630 lport->tt.disc_stop_final(lport);
1f6ff364 631 lport->tt.exch_mgr_reset(lport, 0, 0);
42e9a92f
RL
632 return 0;
633}
634EXPORT_SYMBOL(fc_lport_destroy);
635
636/**
3a3b42bf
RL
637 * fc_set_mfs() - Set the maximum frame size for a local port
638 * @lport: The local port to set the MFS for
639 * @mfs: The new MFS
34f42a07 640 */
42e9a92f
RL
641int fc_set_mfs(struct fc_lport *lport, u32 mfs)
642{
643 unsigned int old_mfs;
644 int rc = -EINVAL;
645
646 mutex_lock(&lport->lp_mutex);
647
648 old_mfs = lport->mfs;
649
650 if (mfs >= FC_MIN_MAX_FRAME) {
651 mfs &= ~3;
652 if (mfs > FC_MAX_FRAME)
653 mfs = FC_MAX_FRAME;
654 mfs -= sizeof(struct fc_frame_header);
655 lport->mfs = mfs;
656 rc = 0;
657 }
658
659 if (!rc && mfs < old_mfs)
660 fc_lport_enter_reset(lport);
661
662 mutex_unlock(&lport->lp_mutex);
663
664 return rc;
665}
666EXPORT_SYMBOL(fc_set_mfs);
667
668/**
34f42a07 669 * fc_lport_disc_callback() - Callback for discovery events
3a3b42bf 670 * @lport: The local port receiving the event
42e9a92f
RL
671 * @event: The discovery event
672 */
673void fc_lport_disc_callback(struct fc_lport *lport, enum fc_disc_event event)
674{
675 switch (event) {
676 case DISC_EV_SUCCESS:
7414705e 677 FC_LPORT_DBG(lport, "Discovery succeeded\n");
42e9a92f
RL
678 break;
679 case DISC_EV_FAILED:
e6d8a1b0 680 printk(KERN_ERR "host%d: libfc: "
ce8b5df0 681 "Discovery failed for port (%6.6x)\n",
7b2787ec 682 lport->host->host_no, lport->port_id);
42e9a92f
RL
683 mutex_lock(&lport->lp_mutex);
684 fc_lport_enter_reset(lport);
685 mutex_unlock(&lport->lp_mutex);
686 break;
687 case DISC_EV_NONE:
688 WARN_ON(1);
689 break;
690 }
691}
692
693/**
34f42a07 694 * fc_rport_enter_ready() - Enter the ready state and start discovery
3a3b42bf 695 * @lport: The local port that is ready
42e9a92f
RL
696 *
697 * Locking Note: The lport lock is expected to be held before calling
698 * this routine.
699 */
700static void fc_lport_enter_ready(struct fc_lport *lport)
701{
7414705e
RL
702 FC_LPORT_DBG(lport, "Entered READY from state %s\n",
703 fc_lport_state(lport));
42e9a92f
RL
704
705 fc_lport_state_enter(lport, LPORT_ST_READY);
8faecddb
CL
706 if (lport->vport)
707 fc_vport_set_state(lport->vport, FC_VPORT_ACTIVE);
708 fc_vports_linkchange(lport);
42e9a92f 709
3a3b42bf 710 if (!lport->ptp_rdata)
29d898e9 711 lport->tt.disc_start(fc_lport_disc_callback, lport);
42e9a92f
RL
712}
713
093bb6a2
JE
714/**
715 * fc_lport_set_port_id() - set the local port Port ID
716 * @lport: The local port which will have its Port ID set.
717 * @port_id: The new port ID.
718 * @fp: The frame containing the incoming request, or NULL.
719 *
720 * Locking Note: The lport lock is expected to be held before calling
721 * this function.
722 */
723static void fc_lport_set_port_id(struct fc_lport *lport, u32 port_id,
724 struct fc_frame *fp)
725{
726 if (port_id)
ce8b5df0 727 printk(KERN_INFO "host%d: Assigned Port ID %6.6x\n",
093bb6a2
JE
728 lport->host->host_no, port_id);
729
7b2787ec
RL
730 lport->port_id = port_id;
731
732 /* Update the fc_host */
093bb6a2 733 fc_host_port_id(lport->host) = port_id;
7b2787ec 734
093bb6a2
JE
735 if (lport->tt.lport_set_port_id)
736 lport->tt.lport_set_port_id(lport, port_id, fp);
737}
738
3726f358
JE
739/**
740 * fc_lport_set_port_id() - set the local port Port ID for point-to-multipoint
741 * @lport: The local port which will have its Port ID set.
742 * @port_id: The new port ID.
743 *
744 * Called by the lower-level driver when transport sets the local port_id.
745 * This is used in VN_port to VN_port mode for FCoE, and causes FLOGI and
746 * discovery to be skipped.
747 */
748void fc_lport_set_local_id(struct fc_lport *lport, u32 port_id)
749{
750 mutex_lock(&lport->lp_mutex);
751
752 fc_lport_set_port_id(lport, port_id, NULL);
753
754 switch (lport->state) {
755 case LPORT_ST_RESET:
756 case LPORT_ST_FLOGI:
757 if (port_id)
758 fc_lport_enter_ready(lport);
759 break;
760 default:
761 break;
762 }
763 mutex_unlock(&lport->lp_mutex);
764}
765EXPORT_SYMBOL(fc_lport_set_local_id);
766
42e9a92f 767/**
34f42a07 768 * fc_lport_recv_flogi_req() - Receive a FLOGI request
3a3b42bf 769 * @lport: The local port that recieved the request
92261156 770 * @rx_fp: The FLOGI frame
42e9a92f
RL
771 *
772 * A received FLOGI request indicates a point-to-point connection.
773 * Accept it with the common service parameters indicating our N port.
774 * Set up to do a PLOGI if we have the higher-number WWPN.
775 *
1b69bc06 776 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
777 * this function.
778 */
92261156
JE
779static void fc_lport_recv_flogi_req(struct fc_lport *lport,
780 struct fc_frame *rx_fp)
42e9a92f
RL
781{
782 struct fc_frame *fp;
24f089e2 783 struct fc_frame_header *fh;
42e9a92f
RL
784 struct fc_els_flogi *flp;
785 struct fc_els_flogi *new_flp;
786 u64 remote_wwpn;
787 u32 remote_fid;
788 u32 local_fid;
42e9a92f 789
7414705e
RL
790 FC_LPORT_DBG(lport, "Received FLOGI request while in state %s\n",
791 fc_lport_state(lport));
42e9a92f 792
251748a9 793 remote_fid = fc_frame_sid(rx_fp);
42e9a92f
RL
794 flp = fc_frame_payload_get(rx_fp, sizeof(*flp));
795 if (!flp)
796 goto out;
797 remote_wwpn = get_unaligned_be64(&flp->fl_wwpn);
798 if (remote_wwpn == lport->wwpn) {
e6d8a1b0 799 printk(KERN_WARNING "host%d: libfc: Received FLOGI from port "
9f8f3aa6 800 "with same WWPN %16.16llx\n",
e6d8a1b0 801 lport->host->host_no, remote_wwpn);
42e9a92f
RL
802 goto out;
803 }
9f8f3aa6 804 FC_LPORT_DBG(lport, "FLOGI from port WWPN %16.16llx\n", remote_wwpn);
42e9a92f
RL
805
806 /*
807 * XXX what is the right thing to do for FIDs?
808 * The originator might expect our S_ID to be 0xfffffe.
809 * But if so, both of us could end up with the same FID.
810 */
811 local_fid = FC_LOCAL_PTP_FID_LO;
812 if (remote_wwpn < lport->wwpn) {
813 local_fid = FC_LOCAL_PTP_FID_HI;
814 if (!remote_fid || remote_fid == local_fid)
815 remote_fid = FC_LOCAL_PTP_FID_LO;
816 } else if (!remote_fid) {
817 remote_fid = FC_LOCAL_PTP_FID_HI;
818 }
819
093bb6a2 820 fc_lport_set_port_id(lport, local_fid, rx_fp);
42e9a92f
RL
821
822 fp = fc_frame_alloc(lport, sizeof(*flp));
823 if (fp) {
42e9a92f
RL
824 new_flp = fc_frame_payload_get(fp, sizeof(*flp));
825 fc_lport_flogi_fill(lport, new_flp, ELS_FLOGI);
826 new_flp->fl_cmd = (u8) ELS_LS_ACC;
827
828 /*
829 * Send the response. If this fails, the originator should
830 * repeat the sequence.
831 */
24f089e2
JE
832 fc_fill_reply_hdr(fp, rx_fp, FC_RCTL_ELS_REP, 0);
833 fh = fc_frame_header_get(fp);
834 hton24(fh->fh_s_id, local_fid);
835 hton24(fh->fh_d_id, remote_fid);
836 lport->tt.frame_send(lport, fp);
42e9a92f
RL
837
838 } else {
839 fc_lport_error(lport, fp);
840 }
841 fc_lport_ptp_setup(lport, remote_fid, remote_wwpn,
842 get_unaligned_be64(&flp->fl_wwnn));
42e9a92f 843out:
42e9a92f
RL
844 fc_frame_free(rx_fp);
845}
846
847/**
34f42a07 848 * fc_lport_recv_req() - The generic lport request handler
3a3b42bf 849 * @lport: The local port that received the request
3a3b42bf 850 * @fp: The request frame
42e9a92f
RL
851 *
852 * This function will see if the lport handles the request or
853 * if an rport should handle the request.
854 *
855 * Locking Note: This function should not be called with the lport
856 * lock held becuase it will grab the lock.
857 */
92261156 858static void fc_lport_recv_req(struct fc_lport *lport, struct fc_frame *fp)
42e9a92f
RL
859{
860 struct fc_frame_header *fh = fc_frame_header_get(fp);
92261156 861 void (*recv)(struct fc_lport *, struct fc_frame *);
42e9a92f
RL
862
863 mutex_lock(&lport->lp_mutex);
864
865 /*
866 * Handle special ELS cases like FLOGI, LOGO, and
867 * RSCN here. These don't require a session.
868 * Even if we had a session, it might not be ready.
869 */
e9ba8b42
JE
870 if (!lport->link_up)
871 fc_frame_free(fp);
872 else if (fh->fh_type == FC_TYPE_ELS &&
873 fh->fh_r_ctl == FC_RCTL_ELS_REQ) {
42e9a92f
RL
874 /*
875 * Check opcode.
876 */
131203a1 877 recv = lport->tt.rport_recv_req;
42e9a92f
RL
878 switch (fc_frame_payload_op(fp)) {
879 case ELS_FLOGI:
a7b12a27
JE
880 if (!lport->point_to_multipoint)
881 recv = fc_lport_recv_flogi_req;
42e9a92f
RL
882 break;
883 case ELS_LOGO:
251748a9 884 if (fc_frame_sid(fp) == FC_FID_FLOGI)
42e9a92f
RL
885 recv = fc_lport_recv_logo_req;
886 break;
887 case ELS_RSCN:
888 recv = lport->tt.disc_recv_req;
889 break;
890 case ELS_ECHO:
891 recv = fc_lport_recv_echo_req;
892 break;
893 case ELS_RLIR:
894 recv = fc_lport_recv_rlir_req;
895 break;
896 case ELS_RNID:
897 recv = fc_lport_recv_rnid_req;
898 break;
42e9a92f
RL
899 }
900
92261156 901 recv(lport, fp);
42e9a92f 902 } else {
7414705e
RL
903 FC_LPORT_DBG(lport, "dropping invalid frame (eof %x)\n",
904 fr_eof(fp));
42e9a92f
RL
905 fc_frame_free(fp);
906 }
907 mutex_unlock(&lport->lp_mutex);
42e9a92f
RL
908}
909
910/**
3a3b42bf
RL
911 * fc_lport_reset() - Reset a local port
912 * @lport: The local port which should be reset
42e9a92f
RL
913 *
914 * Locking Note: This functions should not be called with the
915 * lport lock held.
916 */
917int fc_lport_reset(struct fc_lport *lport)
918{
f7db2c15 919 cancel_delayed_work_sync(&lport->retry_work);
42e9a92f
RL
920 mutex_lock(&lport->lp_mutex);
921 fc_lport_enter_reset(lport);
922 mutex_unlock(&lport->lp_mutex);
923 return 0;
924}
925EXPORT_SYMBOL(fc_lport_reset);
926
927/**
3a3b42bf
RL
928 * fc_lport_reset_locked() - Reset the local port w/ the lport lock held
929 * @lport: The local port to be reset
42e9a92f
RL
930 *
931 * Locking Note: The lport lock is expected to be held before calling
932 * this routine.
933 */
1190d925 934static void fc_lport_reset_locked(struct fc_lport *lport)
42e9a92f 935{
3a3b42bf
RL
936 if (lport->dns_rdata)
937 lport->tt.rport_logoff(lport->dns_rdata);
42e9a92f 938
2f2ac4a0
JE
939 if (lport->ptp_rdata) {
940 lport->tt.rport_logoff(lport->ptp_rdata);
941 kref_put(&lport->ptp_rdata->kref, lport->tt.rport_destroy);
942 lport->ptp_rdata = NULL;
943 }
42e9a92f
RL
944
945 lport->tt.disc_stop(lport);
946
1f6ff364 947 lport->tt.exch_mgr_reset(lport, 0, 0);
42e9a92f 948 fc_host_fabric_name(lport->host) = 0;
093bb6a2 949
3726f358 950 if (lport->port_id && (!lport->point_to_multipoint || !lport->link_up))
093bb6a2 951 fc_lport_set_port_id(lport, 0, NULL);
1190d925 952}
42e9a92f 953
1190d925
JE
954/**
955 * fc_lport_enter_reset() - Reset the local port
3a3b42bf 956 * @lport: The local port to be reset
1190d925
JE
957 *
958 * Locking Note: The lport lock is expected to be held before calling
959 * this routine.
960 */
961static void fc_lport_enter_reset(struct fc_lport *lport)
962{
963 FC_LPORT_DBG(lport, "Entered RESET state from %s state\n",
964 fc_lport_state(lport));
965
55a66d3c
VD
966 if (lport->state == LPORT_ST_DISABLED || lport->state == LPORT_ST_LOGO)
967 return;
968
8faecddb
CL
969 if (lport->vport) {
970 if (lport->link_up)
971 fc_vport_set_state(lport->vport, FC_VPORT_INITIALIZING);
972 else
973 fc_vport_set_state(lport->vport, FC_VPORT_LINKDOWN);
974 }
1190d925 975 fc_lport_state_enter(lport, LPORT_ST_RESET);
8faecddb 976 fc_vports_linkchange(lport);
1190d925 977 fc_lport_reset_locked(lport);
bc0e17f6 978 if (lport->link_up)
42e9a92f
RL
979 fc_lport_enter_flogi(lport);
980}
981
1190d925 982/**
3a3b42bf
RL
983 * fc_lport_enter_disabled() - Disable the local port
984 * @lport: The local port to be reset
1190d925
JE
985 *
986 * Locking Note: The lport lock is expected to be held before calling
987 * this routine.
988 */
989static void fc_lport_enter_disabled(struct fc_lport *lport)
990{
991 FC_LPORT_DBG(lport, "Entered disabled state from %s state\n",
992 fc_lport_state(lport));
993
994 fc_lport_state_enter(lport, LPORT_ST_DISABLED);
8faecddb 995 fc_vports_linkchange(lport);
1190d925
JE
996 fc_lport_reset_locked(lport);
997}
998
42e9a92f 999/**
34f42a07 1000 * fc_lport_error() - Handler for any errors
3a3b42bf
RL
1001 * @lport: The local port that the error was on
1002 * @fp: The error code encoded in a frame pointer
42e9a92f
RL
1003 *
1004 * If the error was caused by a resource allocation failure
1005 * then wait for half a second and retry, otherwise retry
1006 * after the e_d_tov time.
1007 */
1008static void fc_lport_error(struct fc_lport *lport, struct fc_frame *fp)
1009{
1010 unsigned long delay = 0;
7414705e
RL
1011 FC_LPORT_DBG(lport, "Error %ld in state %s, retries %d\n",
1012 PTR_ERR(fp), fc_lport_state(lport),
1013 lport->retry_count);
42e9a92f 1014
7f985231
BPG
1015 if (PTR_ERR(fp) == -FC_EX_CLOSED)
1016 return;
1017
1018 /*
1019 * Memory allocation failure, or the exchange timed out
1020 * or we received LS_RJT.
1021 * Retry after delay
1022 */
1023 if (lport->retry_count < lport->max_retry_count) {
1024 lport->retry_count++;
1025 if (!fp)
1026 delay = msecs_to_jiffies(500);
1027 else
1028 delay = msecs_to_jiffies(lport->e_d_tov);
1029
1030 schedule_delayed_work(&lport->retry_work, delay);
1031 } else
1032 fc_lport_enter_reset(lport);
42e9a92f
RL
1033}
1034
1035/**
7cccc157 1036 * fc_lport_ns_resp() - Handle response to a name server
3a3b42bf
RL
1037 * registration exchange
1038 * @sp: current sequence in exchange
1039 * @fp: response frame
42e9a92f
RL
1040 * @lp_arg: Fibre Channel host port instance
1041 *
1042 * Locking Note: This function will be called without the lport lock
3a3b42bf 1043 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1044 * and then unlock the lport.
1045 */
7cccc157
CL
1046static void fc_lport_ns_resp(struct fc_seq *sp, struct fc_frame *fp,
1047 void *lp_arg)
42e9a92f
RL
1048{
1049 struct fc_lport *lport = lp_arg;
1050 struct fc_frame_header *fh;
1051 struct fc_ct_hdr *ct;
1052
7cccc157 1053 FC_LPORT_DBG(lport, "Received a ns %s\n", fc_els_resp_type(fp));
f657d299 1054
42e9a92f
RL
1055 if (fp == ERR_PTR(-FC_EX_CLOSED))
1056 return;
1057
1058 mutex_lock(&lport->lp_mutex);
1059
ab593b18 1060 if (lport->state < LPORT_ST_RNN_ID || lport->state > LPORT_ST_RFF_ID) {
7cccc157 1061 FC_LPORT_DBG(lport, "Received a name server response, "
3a3b42bf 1062 "but in state %s\n", fc_lport_state(lport));
76f6804e
AJ
1063 if (IS_ERR(fp))
1064 goto err;
42e9a92f
RL
1065 goto out;
1066 }
1067
76f6804e
AJ
1068 if (IS_ERR(fp)) {
1069 fc_lport_error(lport, fp);
1070 goto err;
1071 }
1072
42e9a92f
RL
1073 fh = fc_frame_header_get(fp);
1074 ct = fc_frame_payload_get(fp, sizeof(*ct));
1075
1076 if (fh && ct && fh->fh_type == FC_TYPE_CT &&
1077 ct->ct_fs_type == FC_FST_DIR &&
1078 ct->ct_fs_subtype == FC_NS_SUBTYPE &&
1079 ntohs(ct->ct_cmd) == FC_FS_ACC)
7cccc157
CL
1080 switch (lport->state) {
1081 case LPORT_ST_RNN_ID:
c914f7d1 1082 fc_lport_enter_ns(lport, LPORT_ST_RSNN_NN);
7cccc157
CL
1083 break;
1084 case LPORT_ST_RSNN_NN:
c914f7d1 1085 fc_lport_enter_ns(lport, LPORT_ST_RSPN_ID);
7cccc157
CL
1086 break;
1087 case LPORT_ST_RSPN_ID:
c914f7d1 1088 fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
7cccc157
CL
1089 break;
1090 case LPORT_ST_RFT_ID:
ab593b18
JE
1091 fc_lport_enter_ns(lport, LPORT_ST_RFF_ID);
1092 break;
1093 case LPORT_ST_RFF_ID:
7cccc157
CL
1094 fc_lport_enter_scr(lport);
1095 break;
1096 default:
1097 /* should have already been caught by state checks */
1098 break;
1099 }
c9c7bd7a
CL
1100 else
1101 fc_lport_error(lport, fp);
c9c7bd7a
CL
1102out:
1103 fc_frame_free(fp);
1104err:
1105 mutex_unlock(&lport->lp_mutex);
1106}
1107
42e9a92f 1108/**
34f42a07 1109 * fc_lport_scr_resp() - Handle response to State Change Register (SCR) request
3a3b42bf
RL
1110 * @sp: current sequence in SCR exchange
1111 * @fp: response frame
42e9a92f
RL
1112 * @lp_arg: Fibre Channel lport port instance that sent the registration request
1113 *
1114 * Locking Note: This function will be called without the lport lock
1115 * held, but it will lock, call an _enter_* function or fc_lport_error
1116 * and then unlock the lport.
1117 */
1118static void fc_lport_scr_resp(struct fc_seq *sp, struct fc_frame *fp,
1119 void *lp_arg)
1120{
1121 struct fc_lport *lport = lp_arg;
1122 u8 op;
1123
f657d299
JE
1124 FC_LPORT_DBG(lport, "Received a SCR %s\n", fc_els_resp_type(fp));
1125
42e9a92f
RL
1126 if (fp == ERR_PTR(-FC_EX_CLOSED))
1127 return;
1128
1129 mutex_lock(&lport->lp_mutex);
1130
42e9a92f 1131 if (lport->state != LPORT_ST_SCR) {
7414705e
RL
1132 FC_LPORT_DBG(lport, "Received a SCR response, but in state "
1133 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1134 if (IS_ERR(fp))
1135 goto err;
42e9a92f
RL
1136 goto out;
1137 }
1138
76f6804e
AJ
1139 if (IS_ERR(fp)) {
1140 fc_lport_error(lport, fp);
1141 goto err;
1142 }
1143
42e9a92f
RL
1144 op = fc_frame_payload_op(fp);
1145 if (op == ELS_LS_ACC)
1146 fc_lport_enter_ready(lport);
1147 else
1148 fc_lport_error(lport, fp);
1149
1150out:
1151 fc_frame_free(fp);
1152err:
1153 mutex_unlock(&lport->lp_mutex);
1154}
1155
1156/**
3a3b42bf
RL
1157 * fc_lport_enter_scr() - Send a SCR (State Change Register) request
1158 * @lport: The local port to register for state changes
42e9a92f
RL
1159 *
1160 * Locking Note: The lport lock is expected to be held before calling
1161 * this routine.
1162 */
1163static void fc_lport_enter_scr(struct fc_lport *lport)
1164{
1165 struct fc_frame *fp;
1166
7414705e
RL
1167 FC_LPORT_DBG(lport, "Entered SCR state from %s state\n",
1168 fc_lport_state(lport));
42e9a92f
RL
1169
1170 fc_lport_state_enter(lport, LPORT_ST_SCR);
1171
1172 fp = fc_frame_alloc(lport, sizeof(struct fc_els_scr));
1173 if (!fp) {
1174 fc_lport_error(lport, fp);
1175 return;
1176 }
1177
a46f327a 1178 if (!lport->tt.elsct_send(lport, FC_FID_FCTRL, fp, ELS_SCR,
b94f8951
JE
1179 fc_lport_scr_resp, lport,
1180 2 * lport->r_a_tov))
8f550f93 1181 fc_lport_error(lport, NULL);
42e9a92f
RL
1182}
1183
1184/**
c914f7d1 1185 * fc_lport_enter_ns() - register some object with the name server
42e9a92f
RL
1186 * @lport: Fibre Channel local port to register
1187 *
1188 * Locking Note: The lport lock is expected to be held before calling
1189 * this routine.
1190 */
c914f7d1 1191static void fc_lport_enter_ns(struct fc_lport *lport, enum fc_lport_state state)
c9866a54
CL
1192{
1193 struct fc_frame *fp;
c914f7d1
CL
1194 enum fc_ns_req cmd;
1195 int size = sizeof(struct fc_ct_hdr);
c9866a54
CL
1196 size_t len;
1197
c914f7d1
CL
1198 FC_LPORT_DBG(lport, "Entered %s state from %s state\n",
1199 fc_lport_state_names[state],
c9866a54
CL
1200 fc_lport_state(lport));
1201
c914f7d1 1202 fc_lport_state_enter(lport, state);
c9866a54 1203
c914f7d1
CL
1204 switch (state) {
1205 case LPORT_ST_RNN_ID:
1206 cmd = FC_NS_RNN_ID;
1207 size += sizeof(struct fc_ns_rn_id);
1208 break;
1209 case LPORT_ST_RSNN_NN:
1210 len = strnlen(fc_host_symbolic_name(lport->host), 255);
1211 /* if there is no symbolic name, skip to RFT_ID */
1212 if (!len)
1213 return fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
1214 cmd = FC_NS_RSNN_NN;
1215 size += sizeof(struct fc_ns_rsnn) + len;
1216 break;
1217 case LPORT_ST_RSPN_ID:
1218 len = strnlen(fc_host_symbolic_name(lport->host), 255);
1219 /* if there is no symbolic name, skip to RFT_ID */
1220 if (!len)
1221 return fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
1222 cmd = FC_NS_RSPN_ID;
1223 size += sizeof(struct fc_ns_rspn) + len;
1224 break;
1225 case LPORT_ST_RFT_ID:
1226 cmd = FC_NS_RFT_ID;
1227 size += sizeof(struct fc_ns_rft);
1228 break;
ab593b18
JE
1229 case LPORT_ST_RFF_ID:
1230 cmd = FC_NS_RFF_ID;
1231 size += sizeof(struct fc_ns_rff_id);
1232 break;
c914f7d1
CL
1233 default:
1234 fc_lport_error(lport, NULL);
5baa17c3
CL
1235 return;
1236 }
1237
c914f7d1 1238 fp = fc_frame_alloc(lport, size);
c9c7bd7a
CL
1239 if (!fp) {
1240 fc_lport_error(lport, fp);
1241 return;
1242 }
1243
c914f7d1 1244 if (!lport->tt.elsct_send(lport, FC_FID_DIR_SERV, fp, cmd,
7cccc157 1245 fc_lport_ns_resp,
b94f8951 1246 lport, 3 * lport->r_a_tov))
c9c7bd7a
CL
1247 fc_lport_error(lport, fp);
1248}
1249
42e9a92f
RL
1250static struct fc_rport_operations fc_lport_rport_ops = {
1251 .event_callback = fc_lport_rport_callback,
1252};
1253
1254/**
3a3b42bf
RL
1255 * fc_rport_enter_dns() - Create a fc_rport for the name server
1256 * @lport: The local port requesting a remote port for the name server
42e9a92f
RL
1257 *
1258 * Locking Note: The lport lock is expected to be held before calling
1259 * this routine.
1260 */
1261static void fc_lport_enter_dns(struct fc_lport *lport)
1262{
ab28f1fd 1263 struct fc_rport_priv *rdata;
42e9a92f 1264
7414705e
RL
1265 FC_LPORT_DBG(lport, "Entered DNS state from %s state\n",
1266 fc_lport_state(lport));
42e9a92f
RL
1267
1268 fc_lport_state_enter(lport, LPORT_ST_DNS);
1269
48f00902 1270 mutex_lock(&lport->disc.disc_mutex);
9737e6a7 1271 rdata = lport->tt.rport_create(lport, FC_FID_DIR_SERV);
48f00902 1272 mutex_unlock(&lport->disc.disc_mutex);
9fb9d328 1273 if (!rdata)
42e9a92f
RL
1274 goto err;
1275
42e9a92f 1276 rdata->ops = &fc_lport_rport_ops;
9fb9d328 1277 lport->tt.rport_login(rdata);
42e9a92f
RL
1278 return;
1279
1280err:
1281 fc_lport_error(lport, NULL);
1282}
1283
1284/**
3a3b42bf
RL
1285 * fc_lport_timeout() - Handler for the retry_work timer
1286 * @work: The work struct of the local port
42e9a92f
RL
1287 */
1288static void fc_lport_timeout(struct work_struct *work)
1289{
1290 struct fc_lport *lport =
1291 container_of(work, struct fc_lport,
1292 retry_work.work);
1293
1294 mutex_lock(&lport->lp_mutex);
1295
1296 switch (lport->state) {
b1d9fd55 1297 case LPORT_ST_DISABLED:
22655ac2
JE
1298 WARN_ON(1);
1299 break;
42e9a92f 1300 case LPORT_ST_READY:
42e9a92f
RL
1301 WARN_ON(1);
1302 break;
22655ac2
JE
1303 case LPORT_ST_RESET:
1304 break;
42e9a92f
RL
1305 case LPORT_ST_FLOGI:
1306 fc_lport_enter_flogi(lport);
1307 break;
1308 case LPORT_ST_DNS:
1309 fc_lport_enter_dns(lport);
1310 break;
c9c7bd7a 1311 case LPORT_ST_RNN_ID:
5baa17c3 1312 case LPORT_ST_RSNN_NN:
c9866a54 1313 case LPORT_ST_RSPN_ID:
42e9a92f 1314 case LPORT_ST_RFT_ID:
ab593b18 1315 case LPORT_ST_RFF_ID:
c914f7d1 1316 fc_lport_enter_ns(lport, lport->state);
42e9a92f
RL
1317 break;
1318 case LPORT_ST_SCR:
1319 fc_lport_enter_scr(lport);
1320 break;
1321 case LPORT_ST_LOGO:
1322 fc_lport_enter_logo(lport);
1323 break;
1324 }
1325
1326 mutex_unlock(&lport->lp_mutex);
1327}
1328
1329/**
34f42a07 1330 * fc_lport_logo_resp() - Handle response to LOGO request
3a3b42bf
RL
1331 * @sp: The sequence that the LOGO was on
1332 * @fp: The LOGO frame
1333 * @lp_arg: The lport port that received the LOGO request
42e9a92f
RL
1334 *
1335 * Locking Note: This function will be called without the lport lock
3a3b42bf 1336 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1337 * and then unlock the lport.
1338 */
11b56188 1339void fc_lport_logo_resp(struct fc_seq *sp, struct fc_frame *fp,
3a3b42bf 1340 void *lp_arg)
42e9a92f
RL
1341{
1342 struct fc_lport *lport = lp_arg;
1343 u8 op;
1344
f657d299
JE
1345 FC_LPORT_DBG(lport, "Received a LOGO %s\n", fc_els_resp_type(fp));
1346
42e9a92f
RL
1347 if (fp == ERR_PTR(-FC_EX_CLOSED))
1348 return;
1349
1350 mutex_lock(&lport->lp_mutex);
1351
42e9a92f 1352 if (lport->state != LPORT_ST_LOGO) {
7414705e
RL
1353 FC_LPORT_DBG(lport, "Received a LOGO response, but in state "
1354 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1355 if (IS_ERR(fp))
1356 goto err;
42e9a92f
RL
1357 goto out;
1358 }
1359
76f6804e
AJ
1360 if (IS_ERR(fp)) {
1361 fc_lport_error(lport, fp);
1362 goto err;
1363 }
1364
42e9a92f
RL
1365 op = fc_frame_payload_op(fp);
1366 if (op == ELS_LS_ACC)
1190d925 1367 fc_lport_enter_disabled(lport);
42e9a92f
RL
1368 else
1369 fc_lport_error(lport, fp);
1370
1371out:
1372 fc_frame_free(fp);
1373err:
1374 mutex_unlock(&lport->lp_mutex);
1375}
11b56188 1376EXPORT_SYMBOL(fc_lport_logo_resp);
42e9a92f
RL
1377
1378/**
34f42a07 1379 * fc_rport_enter_logo() - Logout of the fabric
3a3b42bf 1380 * @lport: The local port to be logged out
42e9a92f
RL
1381 *
1382 * Locking Note: The lport lock is expected to be held before calling
1383 * this routine.
1384 */
1385static void fc_lport_enter_logo(struct fc_lport *lport)
1386{
1387 struct fc_frame *fp;
1388 struct fc_els_logo *logo;
1389
7414705e
RL
1390 FC_LPORT_DBG(lport, "Entered LOGO state from %s state\n",
1391 fc_lport_state(lport));
42e9a92f
RL
1392
1393 fc_lport_state_enter(lport, LPORT_ST_LOGO);
8faecddb 1394 fc_vports_linkchange(lport);
42e9a92f 1395
42e9a92f
RL
1396 fp = fc_frame_alloc(lport, sizeof(*logo));
1397 if (!fp) {
1398 fc_lport_error(lport, fp);
1399 return;
1400 }
1401
a46f327a 1402 if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp, ELS_LOGO,
b94f8951
JE
1403 fc_lport_logo_resp, lport,
1404 2 * lport->r_a_tov))
8f550f93 1405 fc_lport_error(lport, NULL);
42e9a92f
RL
1406}
1407
1408/**
34f42a07 1409 * fc_lport_flogi_resp() - Handle response to FLOGI request
3a3b42bf
RL
1410 * @sp: The sequence that the FLOGI was on
1411 * @fp: The FLOGI response frame
1412 * @lp_arg: The lport port that received the FLOGI response
42e9a92f
RL
1413 *
1414 * Locking Note: This function will be called without the lport lock
3a3b42bf 1415 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1416 * and then unlock the lport.
1417 */
11b56188 1418void fc_lport_flogi_resp(struct fc_seq *sp, struct fc_frame *fp,
3a3b42bf 1419 void *lp_arg)
42e9a92f
RL
1420{
1421 struct fc_lport *lport = lp_arg;
42e9a92f
RL
1422 struct fc_els_flogi *flp;
1423 u32 did;
1424 u16 csp_flags;
1425 unsigned int r_a_tov;
1426 unsigned int e_d_tov;
1427 u16 mfs;
1428
f657d299
JE
1429 FC_LPORT_DBG(lport, "Received a FLOGI %s\n", fc_els_resp_type(fp));
1430
42e9a92f
RL
1431 if (fp == ERR_PTR(-FC_EX_CLOSED))
1432 return;
1433
1434 mutex_lock(&lport->lp_mutex);
1435
42e9a92f 1436 if (lport->state != LPORT_ST_FLOGI) {
7414705e
RL
1437 FC_LPORT_DBG(lport, "Received a FLOGI response, but in state "
1438 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1439 if (IS_ERR(fp))
1440 goto err;
42e9a92f
RL
1441 goto out;
1442 }
1443
76f6804e
AJ
1444 if (IS_ERR(fp)) {
1445 fc_lport_error(lport, fp);
1446 goto err;
1447 }
1448
251748a9 1449 did = fc_frame_did(fp);
60a3c4df 1450 if (fc_frame_payload_op(fp) == ELS_LS_ACC && did) {
42e9a92f
RL
1451 flp = fc_frame_payload_get(fp, sizeof(*flp));
1452 if (flp) {
1453 mfs = ntohs(flp->fl_csp.sp_bb_data) &
1454 FC_SP_BB_DATA_MASK;
1455 if (mfs >= FC_SP_MIN_MAX_PAYLOAD &&
1456 mfs < lport->mfs)
1457 lport->mfs = mfs;
1458 csp_flags = ntohs(flp->fl_csp.sp_features);
1459 r_a_tov = ntohl(flp->fl_csp.sp_r_a_tov);
1460 e_d_tov = ntohl(flp->fl_csp.sp_e_d_tov);
1461 if (csp_flags & FC_SP_FT_EDTR)
1462 e_d_tov /= 1000000;
db36c06c
CL
1463
1464 lport->npiv_enabled = !!(csp_flags & FC_SP_FT_NPIV_ACC);
1465
42e9a92f
RL
1466 if ((csp_flags & FC_SP_FT_FPORT) == 0) {
1467 if (e_d_tov > lport->e_d_tov)
1468 lport->e_d_tov = e_d_tov;
1469 lport->r_a_tov = 2 * e_d_tov;
093bb6a2 1470 fc_lport_set_port_id(lport, did, fp);
e6d8a1b0 1471 printk(KERN_INFO "host%d: libfc: "
ce8b5df0 1472 "Port (%6.6x) entered "
e6d8a1b0
JE
1473 "point-to-point mode\n",
1474 lport->host->host_no, did);
251748a9 1475 fc_lport_ptp_setup(lport, fc_frame_sid(fp),
42e9a92f
RL
1476 get_unaligned_be64(
1477 &flp->fl_wwpn),
1478 get_unaligned_be64(
1479 &flp->fl_wwnn));
1480 } else {
1481 lport->e_d_tov = e_d_tov;
1482 lport->r_a_tov = r_a_tov;
1483 fc_host_fabric_name(lport->host) =
1484 get_unaligned_be64(&flp->fl_wwnn);
093bb6a2 1485 fc_lport_set_port_id(lport, did, fp);
42e9a92f
RL
1486 fc_lport_enter_dns(lport);
1487 }
1488 }
60a3c4df
VD
1489 } else {
1490 FC_LPORT_DBG(lport, "FLOGI RJT or bad response\n");
7f985231 1491 fc_lport_error(lport, fp);
60a3c4df 1492 }
42e9a92f
RL
1493
1494out:
1495 fc_frame_free(fp);
1496err:
1497 mutex_unlock(&lport->lp_mutex);
1498}
11b56188 1499EXPORT_SYMBOL(fc_lport_flogi_resp);
42e9a92f
RL
1500
1501/**
34f42a07 1502 * fc_rport_enter_flogi() - Send a FLOGI request to the fabric manager
42e9a92f
RL
1503 * @lport: Fibre Channel local port to be logged in to the fabric
1504 *
1505 * Locking Note: The lport lock is expected to be held before calling
1506 * this routine.
1507 */
1508void fc_lport_enter_flogi(struct fc_lport *lport)
1509{
1510 struct fc_frame *fp;
1511
7414705e
RL
1512 FC_LPORT_DBG(lport, "Entered FLOGI state from %s state\n",
1513 fc_lport_state(lport));
42e9a92f
RL
1514
1515 fc_lport_state_enter(lport, LPORT_ST_FLOGI);
1516
3726f358
JE
1517 if (lport->point_to_multipoint) {
1518 if (lport->port_id)
1519 fc_lport_enter_ready(lport);
1520 return;
1521 }
1522
42e9a92f
RL
1523 fp = fc_frame_alloc(lport, sizeof(struct fc_els_flogi));
1524 if (!fp)
1525 return fc_lport_error(lport, fp);
1526
db36c06c
CL
1527 if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp,
1528 lport->vport ? ELS_FDISC : ELS_FLOGI,
b94f8951
JE
1529 fc_lport_flogi_resp, lport,
1530 lport->vport ? 2 * lport->r_a_tov :
1531 lport->e_d_tov))
8f550f93 1532 fc_lport_error(lport, NULL);
42e9a92f
RL
1533}
1534
3a3b42bf
RL
1535/**
1536 * fc_lport_config() - Configure a fc_lport
1537 * @lport: The local port to be configured
1538 */
42e9a92f
RL
1539int fc_lport_config(struct fc_lport *lport)
1540{
1541 INIT_DELAYED_WORK(&lport->retry_work, fc_lport_timeout);
1542 mutex_init(&lport->lp_mutex);
1543
b1d9fd55 1544 fc_lport_state_enter(lport, LPORT_ST_DISABLED);
42e9a92f
RL
1545
1546 fc_lport_add_fc4_type(lport, FC_TYPE_FCP);
1547 fc_lport_add_fc4_type(lport, FC_TYPE_CT);
1548
1549 return 0;
1550}
1551EXPORT_SYMBOL(fc_lport_config);
1552
3a3b42bf
RL
1553/**
1554 * fc_lport_init() - Initialize the lport layer for a local port
1555 * @lport: The local port to initialize the exchange layer for
1556 */
42e9a92f
RL
1557int fc_lport_init(struct fc_lport *lport)
1558{
1559 if (!lport->tt.lport_recv)
1560 lport->tt.lport_recv = fc_lport_recv_req;
1561
1562 if (!lport->tt.lport_reset)
1563 lport->tt.lport_reset = fc_lport_reset;
1564
1565 fc_host_port_type(lport->host) = FC_PORTTYPE_NPORT;
1566 fc_host_node_name(lport->host) = lport->wwnn;
1567 fc_host_port_name(lport->host) = lport->wwpn;
1568 fc_host_supported_classes(lport->host) = FC_COS_CLASS3;
1569 memset(fc_host_supported_fc4s(lport->host), 0,
1570 sizeof(fc_host_supported_fc4s(lport->host)));
1571 fc_host_supported_fc4s(lport->host)[2] = 1;
1572 fc_host_supported_fc4s(lport->host)[7] = 1;
1573
1574 /* This value is also unchanging */
1575 memset(fc_host_active_fc4s(lport->host), 0,
1576 sizeof(fc_host_active_fc4s(lport->host)));
1577 fc_host_active_fc4s(lport->host)[2] = 1;
1578 fc_host_active_fc4s(lport->host)[7] = 1;
1579 fc_host_maxframe_size(lport->host) = lport->mfs;
1580 fc_host_supported_speeds(lport->host) = 0;
1581 if (lport->link_supported_speeds & FC_PORTSPEED_1GBIT)
1582 fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_1GBIT;
1583 if (lport->link_supported_speeds & FC_PORTSPEED_10GBIT)
1584 fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_10GBIT;
1585
1586 return 0;
1587}
1588EXPORT_SYMBOL(fc_lport_init);
a51ab396
SM
1589
1590/**
3a3b42bf
RL
1591 * fc_lport_bsg_resp() - The common response handler for FC Passthrough requests
1592 * @sp: The sequence for the FC Passthrough response
1593 * @fp: The response frame
1594 * @info_arg: The BSG info that the response is for
a51ab396
SM
1595 */
1596static void fc_lport_bsg_resp(struct fc_seq *sp, struct fc_frame *fp,
1597 void *info_arg)
1598{
1599 struct fc_bsg_info *info = info_arg;
1600 struct fc_bsg_job *job = info->job;
1601 struct fc_lport *lport = info->lport;
1602 struct fc_frame_header *fh;
1603 size_t len;
1604 void *buf;
1605
1606 if (IS_ERR(fp)) {
1607 job->reply->result = (PTR_ERR(fp) == -FC_EX_CLOSED) ?
1608 -ECONNABORTED : -ETIMEDOUT;
1609 job->reply_len = sizeof(uint32_t);
1610 job->state_flags |= FC_RQST_STATE_DONE;
1611 job->job_done(job);
1612 kfree(info);
1613 return;
1614 }
1615
1616 mutex_lock(&lport->lp_mutex);
1617 fh = fc_frame_header_get(fp);
1618 len = fr_len(fp) - sizeof(*fh);
1619 buf = fc_frame_payload_get(fp, 0);
1620
1621 if (fr_sof(fp) == FC_SOF_I3 && !ntohs(fh->fh_seq_cnt)) {
1622 /* Get the response code from the first frame payload */
1623 unsigned short cmd = (info->rsp_code == FC_FS_ACC) ?
1624 ntohs(((struct fc_ct_hdr *)buf)->ct_cmd) :
1625 (unsigned short)fc_frame_payload_op(fp);
1626
1627 /* Save the reply status of the job */
1628 job->reply->reply_data.ctels_reply.status =
1629 (cmd == info->rsp_code) ?
1630 FC_CTELS_STATUS_OK : FC_CTELS_STATUS_REJECT;
1631 }
1632
1633 job->reply->reply_payload_rcv_len +=
1634 fc_copy_buffer_to_sglist(buf, len, info->sg, &info->nents,
1635 &info->offset, KM_BIO_SRC_IRQ, NULL);
1636
1637 if (fr_eof(fp) == FC_EOF_T &&
1638 (ntoh24(fh->fh_f_ctl) & (FC_FC_LAST_SEQ | FC_FC_END_SEQ)) ==
1639 (FC_FC_LAST_SEQ | FC_FC_END_SEQ)) {
1640 if (job->reply->reply_payload_rcv_len >
1641 job->reply_payload.payload_len)
1642 job->reply->reply_payload_rcv_len =
1643 job->reply_payload.payload_len;
1644 job->reply->result = 0;
1645 job->state_flags |= FC_RQST_STATE_DONE;
1646 job->job_done(job);
1647 kfree(info);
1648 }
1649 fc_frame_free(fp);
1650 mutex_unlock(&lport->lp_mutex);
1651}
1652
1653/**
3a3b42bf
RL
1654 * fc_lport_els_request() - Send ELS passthrough request
1655 * @job: The BSG Passthrough job
a51ab396 1656 * @lport: The local port sending the request
3a3b42bf 1657 * @did: The destination port id
a51ab396
SM
1658 *
1659 * Locking Note: The lport lock is expected to be held before calling
1660 * this routine.
1661 */
1662static int fc_lport_els_request(struct fc_bsg_job *job,
1663 struct fc_lport *lport,
1664 u32 did, u32 tov)
1665{
1666 struct fc_bsg_info *info;
1667 struct fc_frame *fp;
1668 struct fc_frame_header *fh;
1669 char *pp;
1670 int len;
1671
70d919fb 1672 fp = fc_frame_alloc(lport, job->request_payload.payload_len);
a51ab396
SM
1673 if (!fp)
1674 return -ENOMEM;
1675
1676 len = job->request_payload.payload_len;
1677 pp = fc_frame_payload_get(fp, len);
1678
1679 sg_copy_to_buffer(job->request_payload.sg_list,
1680 job->request_payload.sg_cnt,
1681 pp, len);
1682
1683 fh = fc_frame_header_get(fp);
1684 fh->fh_r_ctl = FC_RCTL_ELS_REQ;
1685 hton24(fh->fh_d_id, did);
7b2787ec 1686 hton24(fh->fh_s_id, lport->port_id);
a51ab396 1687 fh->fh_type = FC_TYPE_ELS;
24f089e2 1688 hton24(fh->fh_f_ctl, FC_FCTL_REQ);
a51ab396
SM
1689 fh->fh_cs_ctl = 0;
1690 fh->fh_df_ctl = 0;
1691 fh->fh_parm_offset = 0;
1692
1693 info = kzalloc(sizeof(struct fc_bsg_info), GFP_KERNEL);
1694 if (!info) {
1695 fc_frame_free(fp);
1696 return -ENOMEM;
1697 }
1698
1699 info->job = job;
1700 info->lport = lport;
1701 info->rsp_code = ELS_LS_ACC;
1702 info->nents = job->reply_payload.sg_cnt;
1703 info->sg = job->reply_payload.sg_list;
1704
1705 if (!lport->tt.exch_seq_send(lport, fp, fc_lport_bsg_resp,
1706 NULL, info, tov))
1707 return -ECOMM;
1708 return 0;
1709}
1710
1711/**
3a3b42bf
RL
1712 * fc_lport_ct_request() - Send CT Passthrough request
1713 * @job: The BSG Passthrough job
a51ab396
SM
1714 * @lport: The local port sending the request
1715 * @did: The destination FC-ID
3a3b42bf 1716 * @tov: The timeout period to wait for the response
a51ab396
SM
1717 *
1718 * Locking Note: The lport lock is expected to be held before calling
1719 * this routine.
1720 */
1721static int fc_lport_ct_request(struct fc_bsg_job *job,
1722 struct fc_lport *lport, u32 did, u32 tov)
1723{
1724 struct fc_bsg_info *info;
1725 struct fc_frame *fp;
1726 struct fc_frame_header *fh;
1727 struct fc_ct_req *ct;
1728 size_t len;
1729
1730 fp = fc_frame_alloc(lport, sizeof(struct fc_ct_hdr) +
1731 job->request_payload.payload_len);
1732 if (!fp)
1733 return -ENOMEM;
1734
1735 len = job->request_payload.payload_len;
1736 ct = fc_frame_payload_get(fp, len);
1737
1738 sg_copy_to_buffer(job->request_payload.sg_list,
1739 job->request_payload.sg_cnt,
1740 ct, len);
1741
1742 fh = fc_frame_header_get(fp);
1743 fh->fh_r_ctl = FC_RCTL_DD_UNSOL_CTL;
1744 hton24(fh->fh_d_id, did);
7b2787ec 1745 hton24(fh->fh_s_id, lport->port_id);
a51ab396 1746 fh->fh_type = FC_TYPE_CT;
24f089e2 1747 hton24(fh->fh_f_ctl, FC_FCTL_REQ);
a51ab396
SM
1748 fh->fh_cs_ctl = 0;
1749 fh->fh_df_ctl = 0;
1750 fh->fh_parm_offset = 0;
1751
1752 info = kzalloc(sizeof(struct fc_bsg_info), GFP_KERNEL);
1753 if (!info) {
1754 fc_frame_free(fp);
1755 return -ENOMEM;
1756 }
1757
1758 info->job = job;
1759 info->lport = lport;
1760 info->rsp_code = FC_FS_ACC;
1761 info->nents = job->reply_payload.sg_cnt;
1762 info->sg = job->reply_payload.sg_list;
1763
1764 if (!lport->tt.exch_seq_send(lport, fp, fc_lport_bsg_resp,
1765 NULL, info, tov))
1766 return -ECOMM;
1767 return 0;
1768}
1769
1770/**
1771 * fc_lport_bsg_request() - The common entry point for sending
3a3b42bf
RL
1772 * FC Passthrough requests
1773 * @job: The BSG passthrough job
a51ab396
SM
1774 */
1775int fc_lport_bsg_request(struct fc_bsg_job *job)
1776{
1777 struct request *rsp = job->req->next_rq;
1778 struct Scsi_Host *shost = job->shost;
1779 struct fc_lport *lport = shost_priv(shost);
1780 struct fc_rport *rport;
1781 struct fc_rport_priv *rdata;
1782 int rc = -EINVAL;
1783 u32 did;
1784
1785 job->reply->reply_payload_rcv_len = 0;
b248df30
HD
1786 if (rsp)
1787 rsp->resid_len = job->reply_payload.payload_len;
a51ab396
SM
1788
1789 mutex_lock(&lport->lp_mutex);
1790
1791 switch (job->request->msgcode) {
1792 case FC_BSG_RPT_ELS:
1793 rport = job->rport;
1794 if (!rport)
1795 break;
1796
1797 rdata = rport->dd_data;
1798 rc = fc_lport_els_request(job, lport, rport->port_id,
1799 rdata->e_d_tov);
1800 break;
1801
1802 case FC_BSG_RPT_CT:
1803 rport = job->rport;
1804 if (!rport)
1805 break;
1806
1807 rdata = rport->dd_data;
1808 rc = fc_lport_ct_request(job, lport, rport->port_id,
1809 rdata->e_d_tov);
1810 break;
1811
1812 case FC_BSG_HST_CT:
1813 did = ntoh24(job->request->rqst_data.h_ct.port_id);
1814 if (did == FC_FID_DIR_SERV)
3a3b42bf 1815 rdata = lport->dns_rdata;
a51ab396
SM
1816 else
1817 rdata = lport->tt.rport_lookup(lport, did);
1818
1819 if (!rdata)
1820 break;
1821
1822 rc = fc_lport_ct_request(job, lport, did, rdata->e_d_tov);
1823 break;
1824
1825 case FC_BSG_HST_ELS_NOLOGIN:
1826 did = ntoh24(job->request->rqst_data.h_els.port_id);
1827 rc = fc_lport_els_request(job, lport, did, lport->e_d_tov);
1828 break;
1829 }
1830
1831 mutex_unlock(&lport->lp_mutex);
1832 return rc;
1833}
1834EXPORT_SYMBOL(fc_lport_bsg_request);