1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
|
#ifndef _QIB_KERNEL_H
#define _QIB_KERNEL_H
/*
* Copyright (c) 2012, 2013 Intel Corporation. All rights reserved.
* Copyright (c) 2006 - 2012 QLogic Corporation. All rights reserved.
* Copyright (c) 2003, 2004, 2005, 2006 PathScale, Inc. All rights reserved.
*
* This software is available to you under a choice of one of two
* licenses. You may choose to be licensed under the terms of the GNU
* General Public License (GPL) Version 2, available from the file
* COPYING in the main directory of this source tree, or the
* OpenIB.org BSD license below:
*
* Redistribution and use in source and binary forms, with or
* without modification, are permitted provided that the following
* conditions are met:
*
* - Redistributions of source code must retain the above
* copyright notice, this list of conditions and the following
* disclaimer.
*
* - Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials
* provided with the distribution.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
/*
* This header file is the base header file for qlogic_ib kernel code
* qib_user.h serves a similar purpose for user code.
*/
#include <linux/interrupt.h>
#include <linux/pci.h>
#include <linux/dma-mapping.h>
#include <linux/mutex.h>
#include <linux/list.h>
#include <linux/scatterlist.h>
#include <linux/slab.h>
#include <linux/io.h>
#include <linux/fs.h>
#include <linux/completion.h>
#include <linux/kref.h>
#include <linux/sched.h>
#include <linux/kthread.h>
#include "qib_common.h"
#include "qib_verbs.h"
/* only s/w major version of QLogic_IB we can handle */
#define QIB_CHIP_VERS_MAJ 2U
/* don't care about this except printing */
#define QIB_CHIP_VERS_MIN 0U
/* The Organization Unique Identifier (Mfg code), and its position in GUID */
#define QIB_OUI 0x001175
#define QIB_OUI_LSB 40
/*
* per driver stats, either not device nor port-specific, or
* summed over all of the devices and ports.
* They are described by name via ipathfs filesystem, so layout
* and number of elements can change without breaking compatibility.
* If members are added or deleted qib_statnames[] in qib_fs.c must
* change to match.
*/
struct qlogic_ib_stats {
__u64 sps_ints; /* number of interrupts handled */
__u64 sps_errints; /* number of error interrupts */
__u64 sps_txerrs; /* tx-related packet errors */
__u64 sps_rcverrs; /* non-crc rcv packet errors */
__u64 sps_hwerrs; /* hardware errors reported (parity, etc.) */
__u64 sps_nopiobufs; /* no pio bufs avail from kernel */
__u64 sps_ctxts; /* number of contexts currently open */
__u64 sps_lenerrs; /* number of kernel packets where RHF != LRH len */
__u64 sps_buffull;
__u64 sps_hdrfull;
};
extern struct qlogic_ib_stats qib_stats;
extern const struct pci_error_handlers qib_pci_err_handler;
extern struct pci_driver qib_driver;
#define QIB_CHIP_SWVERSION QIB_CHIP_VERS_MAJ
/*
* First-cut critierion for "device is active" is
* two thousand dwords combined Tx, Rx traffic per
* 5-second interval. SMA packets are 64 dwords,
* and occur "a few per second", presumably each way.
*/
#define QIB_TRAFFIC_ACTIVE_THRESHOLD (2000)
/*
* Struct used to indicate which errors are logged in each of the
* error-counters that are logged to EEPROM. A counter is incremented
* _once_ (saturating at 255) for each event with any bits set in
* the error or hwerror register masks below.
*/
#define QIB_EEP_LOG_CNT (4)
struct qib_eep_log_mask {
u64 errs_to_log;
u64 hwerrs_to_log;
};
/*
* Below contains all data related to a single context (formerly called port).
*/
#ifdef CONFIG_DEBUG_FS
struct qib_opcode_stats_perctx;
#endif
struct qib_ctxtdata {
void **rcvegrbuf;
dma_addr_t *rcvegrbuf_phys;
/* rcvhdrq base, needs mmap before useful */
void *rcvhdrq;
/* kernel virtual address where hdrqtail is updated */
void *rcvhdrtail_kvaddr;
/*
* temp buffer for expected send setup, allocated at open, instead
* of each setup call
*/
void *tid_pg_list;
/*
* Shared page for kernel to signal user processes that send buffers
* need disarming. The process should call QIB_CMD_DISARM_BUFS
* or QIB_CMD_ACK_EVENT with IPATH_EVENT_DISARM_BUFS set.
*/
unsigned long *user_event_mask;
/* when waiting for rcv or pioavail */
wait_queue_head_t wait;
/*
* rcvegr bufs base, physical, must fit
* in 44 bits so 32 bit programs mmap64 44 bit works)
*/
dma_addr_t rcvegr_phys;
/* mmap of hdrq, must fit in 44 bits */
dma_addr_t rcvhdrq_phys;
dma_addr_t rcvhdrqtailaddr_phys;
/*
* number of opens (including slave sub-contexts) on this instance
* (ignoring forks, dup, etc. for now)
*/
int cnt;
/*
* how much space to leave at start of eager TID entries for
* protocol use, on each TID
*/
/* instead of calculating it */
unsigned ctxt;
/* local node of context */
int node_id;
/* non-zero if ctxt is being shared. */
u16 subctxt_cnt;
/* non-zero if ctxt is being shared. */
u16 subctxt_id;
/* number of eager TID entries. */
u16 rcvegrcnt;
/* index of first eager TID entry. */
u16 rcvegr_tid_base;
/* number of pio bufs for this ctxt (all procs, if shared) */
u32 piocnt;
/* first pio buffer for this ctxt */
u32 pio_base;
/* chip offset of PIO buffers for this ctxt */
u32 piobufs;
/* how many alloc_pages() chunks in rcvegrbuf_pages */
u32 rcvegrbuf_chunks;
/* how many egrbufs per chunk */
u16 rcvegrbufs_perchunk;
/* ilog2 of above */
u16 rcvegrbufs_perchunk_shift;
/* order for rcvegrbuf_pages */
size_t rcvegrbuf_size;
/* rcvhdrq size (for freeing) */
size_t rcvhdrq_size;
/* per-context flags for fileops/intr communication */
unsigned long flag;
/* next expected TID to check when looking for free */
u32 tidcursor;
/* WAIT_RCV that timed out, no interrupt */
u32 rcvwait_to;
/* WAIT_PIO that timed out, no interrupt */
u32 piowait_to;
/* WAIT_RCV already happened, no wait */
u32 rcvnowait;
/* WAIT_PIO already happened, no wait */
u32 pionowait;
/* total number of polled urgent packets */
u32 urgent;
/* saved total number of polled urgent packets for poll edge trigger */
u32 urgent_poll;
/* pid of process using this ctxt */
pid_t pid;
pid_t subpid[QLOGIC_IB_MAX_SUBCTXT];
/* same size as task_struct .comm[], command that opened context */
char comm[16];
/* pkeys set by this use of this ctxt */
u16 pkeys[4];
/* so file ops can get at unit */
struct qib_devdata *dd;
/* so funcs that need physical port can get it easily */
struct qib_pportdata *ppd;
/* A page of memory for rcvhdrhead, rcvegrhead, rcvegrtail * N */
void *subctxt_uregbase;
/* An array of pages for the eager receive buffers * N */
void *subctxt_rcvegrbuf;
/* An array of pages for the eager header queue entries * N */
void *subctxt_rcvhdr_base;
/* The version of the library which opened this ctxt */
u32 userversion;
/* Bitmask of active slaves */
u32 active_slaves;
/* Type of packets or conditions we want to poll for */
u16 poll_type;
/* receive packet sequence counter */
u8 seq_cnt;
u8 redirect_seq_cnt;
/* ctxt rcvhdrq head offset */
u32 head;
/* lookaside fields */
struct qib_qp *lookaside_qp;
u32 lookaside_qpn;
/* QPs waiting for context processing */
struct list_head qp_wait_list;
#ifdef CONFIG_DEBUG_FS
/* verbs stats per CTX */
struct qib_opcode_stats_perctx *opstats;
#endif
};
struct qib_sge_state;
struct qib_sdma_txreq {
int flags;
int sg_count;
dma_addr_t addr;
void (*callback)(struct qib_sdma_txreq *, int);
u16 start_idx; /* sdma private */
u16 next_descq_idx; /* sdma private */
struct list_head list; /* sdma private */
};
struct qib_sdma_desc {
__le64 qw[2];
};
struct qib_verbs_txreq {
struct qib_sdma_txreq txreq;
struct qib_qp *qp;
struct qib_swqe *wqe;
u32 dwords;
u16 hdr_dwords;
u16 hdr_inx;
struct qib_pio_header *align_buf;
struct qib_mregion *mr;
struct qib_sge_state *ss;
};
#define QIB_SDMA_TXREQ_F_USELARGEBUF 0x1
#define QIB_SDMA_TXREQ_F_HEADTOHOST 0x2
#define QIB_SDMA_TXREQ_F_INTREQ 0x4
#define QIB_SDMA_TXREQ_F_FREEBUF 0x8
#define QIB_SDMA_TXREQ_F_FREEDESC 0x10
#define QIB_SDMA_TXREQ_S_OK 0
#define QIB_SDMA_TXREQ_S_SENDERROR 1
#define QIB_SDMA_TXREQ_S_ABORTED 2
#define QIB_SDMA_TXREQ_S_SHUTDOWN 3
/*
* Get/Set IB link-level config parameters for f_get/set_ib_cfg()
* Mostly for MADs that set or query link parameters, also ipath
* config interfaces
*/
#define QIB_IB_CFG_LIDLMC 0 /* LID (LS16b) and Mask (MS16b) */
#define QIB_IB_CFG_LWID_ENB 2 /* allowed Link-width */
#define QIB_IB_CFG_LWID 3 /* currently active Link-width */
#define QIB_IB_CFG_SPD_ENB 4 /* allowed Link speeds */
#define QIB_IB_CFG_SPD 5 /* current Link spd */
#define QIB_IB_CFG_RXPOL_ENB 6 /* Auto-RX-polarity enable */
#define QIB_IB_CFG_LREV_ENB 7 /* Auto-Lane-reversal enable */
#define QIB_IB_CFG_LINKLATENCY 8 /* Link Latency (IB1.2 only) */
#define QIB_IB_CFG_HRTBT 9 /* IB heartbeat off/enable/auto; DDR/QDR only */
#define QIB_IB_CFG_OP_VLS 10 /* operational VLs */
#define QIB_IB_CFG_VL_HIGH_CAP 11 /* num of VL high priority weights */
#define QIB_IB_CFG_VL_LOW_CAP 12 /* num of VL low priority weights */
#define QIB_IB_CFG_OVERRUN_THRESH 13 /* IB overrun threshold */
#define QIB_IB_CFG_PHYERR_THRESH 14 /* IB PHY error threshold */
#define QIB_IB_CFG_LINKDEFAULT 15 /* IB link default (sleep/poll) */
#define QIB_IB_CFG_PKEYS 16 /* update partition keys */
#define QIB_IB_CFG_MTU 17 /* update MTU in IBC */
#define QIB_IB_CFG_LSTATE 18 /* update linkcmd and linkinitcmd in IBC */
#define QIB_IB_CFG_VL_HIGH_LIMIT 19
#define QIB_IB_CFG_PMA_TICKS 20 /* PMA sample tick resolution */
#define QIB_IB_CFG_PORT 21 /* switch port we are connected to */
/*
* for CFG_LSTATE: LINKCMD in upper 16 bits, LINKINITCMD in lower 16
* IB_LINKINITCMD_POLL and SLEEP are also used as set/get values for
* QIB_IB_CFG_LINKDEFAULT cmd
*/
#define IB_LINKCMD_DOWN (0 << 16)
#define IB_LINKCMD_ARMED (1 << 16)
#define IB_LINKCMD_ACTIVE (2 << 16)
#define IB_LINKINITCMD_NOP 0
#define IB_LINKINITCMD_POLL 1
#define IB_LINKINITCMD_SLEEP 2
#define IB_LINKINITCMD_DISABLE 3
/*
* valid states passed to qib_set_linkstate() user call
*/
#define QIB_IB_LINKDOWN 0
#define QIB_IB_LINKARM 1
#define QIB_IB_LINKACTIVE 2
#define QIB_IB_LINKDOWN_ONLY 3
#define QIB_IB_LINKDOWN_SLEEP 4
#define QIB_IB_LINKDOWN_DISABLE 5
/*
* These 7 values (SDR, DDR, and QDR may be ORed for auto-speed
* negotiation) are used for the 3rd argument to path_f_set_ib_cfg
* with cmd QIB_IB_CFG_SPD_ENB, by direct calls or via sysfs. They
* are also the the possible values for qib_link_speed_enabled and active
* The values were chosen to match values used within the IB spec.
*/
#define QIB_IB_SDR 1
#define QIB_IB_DDR 2
#define QIB_IB_QDR 4
#define QIB_DEFAULT_MTU 4096
/* max number of IB ports supported per HCA */
#define QIB_MAX_IB_PORTS 2
/*
* Possible IB config parameters for f_get/set_ib_table()
*/
#define QIB_IB_TBL_VL_HIGH_ARB 1 /* Get/set VL high priority weights */
#define QIB_IB_TBL_VL_LOW_ARB 2 /* Get/set VL low priority weights */
/*
* Possible "operations" for f_rcvctrl(ppd, op, ctxt)
* these are bits so they can be combined, e.g.
* QIB_RCVCTRL_INTRAVAIL_ENB | QIB_RCVCTRL_CTXT_ENB
*/
#define QIB_RCVCTRL_TAILUPD_ENB 0x01
#define QIB_RCVCTRL_TAILUPD_DIS 0x02
#define QIB_RCVCTRL_CTXT_ENB 0x04
#define QIB_RCVCTRL_CTXT_DIS 0x08
#define QIB_RCVCTRL_INTRAVAIL_ENB 0x10
#define QIB_RCVCTRL_INTRAVAIL_DIS 0x20
#define QIB_RCVCTRL_PKEY_ENB 0x40 /* Note, default is enabled */
#define QIB_RCVCTRL_PKEY_DIS 0x80
#define QIB_RCVCTRL_BP_ENB 0x0100
#define QIB_RCVCTRL_BP_DIS 0x0200
#define QIB_RCVCTRL_TIDFLOW_ENB 0x0400
#define QIB_RCVCTRL_TIDFLOW_DIS 0x0800
/*
* Possible "operations" for f_sendctrl(ppd, op, var)
* these are bits so they can be combined, e.g.
* QIB_SENDCTRL_BUFAVAIL_ENB | QIB_SENDCTRL_ENB
* Some operations (e.g. DISARM, ABORT) are known to
* be "one-shot", so do not modify shadow.
*/
#define QIB_SENDCTRL_DISARM (0x1000)
#define QIB_SENDCTRL_DISARM_BUF(bufn) ((bufn) | QIB_SENDCTRL_DISARM)
/* available (0x2000) */
#define QIB_SENDCTRL_AVAIL_DIS (0x4000)
#define QIB_SENDCTRL_AVAIL_ENB (0x8000)
#define QIB_SENDCTRL_AVAIL_BLIP (0x10000)
#define QIB_SENDCTRL_SEND_DIS (0x20000)
#define QIB_SENDCTRL_SEND_ENB (0x40000)
#define QIB_SENDCTRL_FLUSH (0x80000)
#define QIB_SENDCTRL_CLEAR (0x100000)
#define QIB_SENDCTRL_DISARM_ALL (0x200000)
/*
* These are the generic indices for requesting per-port
* counter values via the f_portcntr function. They
* are always returned as 64 bit values, although most
* are 32 bit counters.
*/
/* send-related counters */
#define QIBPORTCNTR_PKTSEND 0U
#define QIBPORTCNTR_WORDSEND 1U
#define QIBPORTCNTR_PSXMITDATA 2U
#define QIBPORTCNTR_PSXMITPKTS 3U
#define QIBPORTCNTR_PSXMITWAIT 4U
#define QIBPORTCNTR_SENDSTALL 5U
/* receive-related counters */
#define QIBPORTCNTR_PKTRCV 6U
#define QIBPORTCNTR_PSRCVDATA 7U
#define QIBPORTCNTR_PSRCVPKTS 8U
#define QIBPORTCNTR_RCVEBP 9U
#define QIBPORTCNTR_RCVOVFL 10U
#define QIBPORTCNTR_WORDRCV 11U
/* IB link related error counters */
#define QIBPORTCNTR_RXLOCALPHYERR 12U
#define QIBPORTCNTR_RXVLERR 13U
#define QIBPORTCNTR_ERRICRC 14U
#define QIBPORTCNTR_ERRVCRC 15U
#define QIBPORTCNTR_ERRLPCRC 16U
#define QIBPORTCNTR_BADFORMAT 17U
#define QIBPORTCNTR_ERR_RLEN 18U
#define QIBPORTCNTR_IBSYMBOLERR 19U
#define QIBPORTCNTR_INVALIDRLEN 20U
#define QIBPORTCNTR_UNSUPVL 21U
#define QIBPORTCNTR_EXCESSBUFOVFL 22U
#define QIBPORTCNTR_ERRLINK 23U
#define QIBPORTCNTR_IBLINKDOWN 24U
#define QIBPORTCNTR_IBLINKERRRECOV 25U
#define QIBPORTCNTR_LLI 26U
/* other error counters */
#define QIBPORTCNTR_RXDROPPKT 27U
#define QIBPORTCNTR_VL15PKTDROP 28U
#define QIBPORTCNTR_ERRPKEY 29U
#define QIBPORTCNTR_KHDROVFL 30U
/* sampling counters (these are actually control registers) */
#define QIBPORTCNTR_PSINTERVAL 31U
#define QIBPORTCNTR_PSSTART 32U
#define QIBPORTCNTR_PSSTAT 33U
/* how often we check for packet activity for "power on hours (in seconds) */
#define ACTIVITY_TIMER 5
#define MAX_NAME_SIZE 64
#ifdef CONFIG_INFINIBAND_QIB_DCA
struct qib_irq_notify;
#endif
struct qib_msix_entry {
struct msix_entry msix;
void *arg;
#ifdef CONFIG_INFINIBAND_QIB_DCA
int dca;
int rcv;
struct qib_irq_notify *notifier;
#endif
char name[MAX_NAME_SIZE];
cpumask_var_t mask;
};
/* Below is an opaque struct. Each chip (device) can maintain
* private data needed for its operation, but not germane to the
* rest of the driver. For convenience, we define another that
* is chip-specific, per-port
*/
struct qib_chip_specific;
struct qib_chipport_specific;
enum qib_sdma_states {
qib_sdma_state_s00_hw_down,
qib_sdma_state_s10_hw_start_up_wait,
qib_sdma_state_s20_idle,
qib_sdma_state_s30_sw_clean_up_wait,
qib_sdma_state_s40_hw_clean_up_wait,
qib_sdma_state_s50_hw_halt_wait,
qib_sdma_state_s99_running,
};
enum qib_sdma_events {
qib_sdma_event_e00_go_hw_down,
qib_sdma_event_e10_go_hw_start,
qib_sdma_event_e20_hw_started,
qib_sdma_event_e30_go_running,
qib_sdma_event_e40_sw_cleaned,
qib_sdma_event_e50_hw_cleaned,
qib_sdma_event_e60_hw_halted,
qib_sdma_event_e70_go_idle,
qib_sdma_event_e7220_err_halted,
qib_sdma_event_e7322_err_halted,
qib_sdma_event_e90_timer_tick,
};
extern char *qib_sdma_state_names[];
extern char *qib_sdma_event_names[];
struct sdma_set_state_action {
unsigned op_enable:1;
unsigned op_intenable:1;
unsigned op_halt:1;
unsigned op_drain:1;
unsigned go_s99_running_tofalse:1;
unsigned go_s99_running_totrue:1;
};
struct qib_sdma_state {
struct kref kref;
struct completion comp;
enum qib_sdma_states current_state;
struct sdma_set_state_action *set_state_action;
unsigned current_op;
unsigned go_s99_running;
unsigned first_sendbuf;
unsigned last_sendbuf; /* really last +1 */
/* debugging/devel */
enum qib_sdma_states previous_state;
unsigned previous_op;
enum qib_sdma_events last_event;
};
struct xmit_wait {
struct timer_list timer;
u64 counter;
u8 flags;
struct cache {
u64 psxmitdata;
u64 psrcvdata;
u64 psxmitpkts;
u64 psrcvpkts;
u64 psxmitwait;
} counter_cache;
};
/*
* The structure below encapsulates data relevant to a physical IB Port.
* Current chips support only one such port, but the separation
* clarifies things a bit. Note that to conform to IB conventions,
* port-numbers are one-based. The first or only port is port1.
*/
struct qib_pportdata {
struct qib_ibport ibport_data;
struct qib_devdata *dd;
struct qib_chippport_specific *cpspec; /* chip-specific per-port */
struct kobject pport_kobj;
struct kobject pport_cc_kobj;
struct kobject sl2vl_kobj;
struct kobject diagc_kobj;
/* GUID for this interface, in network order */
__be64 guid;
/* QIB_POLL, etc. link-state specific flags, per port */
u32 lflags;
/* qib_lflags driver is waiting for */
u32 state_wanted;
spinlock_t lflags_lock;
/* ref count for each pkey */
atomic_t pkeyrefs[4];
/*
* this address is mapped readonly into user processes so they can
* get status cheaply, whenever they want. One qword of status per port
*/
u64 *statusp;
/* SendDMA related entries */
/* read mostly */
struct qib_sdma_desc *sdma_descq;
struct workqueue_struct *qib_wq;
struct qib_sdma_state sdma_state;
dma_addr_t sdma_descq_phys;
volatile __le64 *sdma_head_dma; /* DMA'ed by chip */
dma_addr_t sdma_head_phys;
u16 sdma_descq_cnt;
/* read/write using lock */
spinlock_t sdma_lock ____cacheline_aligned_in_smp;
struct list_head sdma_activelist;
u64 sdma_descq_added;
u64 sdma_descq_removed;
u16 sdma_descq_tail;
u16 sdma_descq_head;
u8 sdma_generation;
struct tasklet_struct sdma_sw_clean_up_task
____cacheline_aligned_in_smp;
wait_queue_head_t state_wait; /* for state_wanted */
/* HoL blocking for SMP replies */
unsigned hol_state;
struct timer_list hol_timer;
/*
* Shadow copies of registers; size indicates read access size.
* Most of them are readonly, but some are write-only register,
* where we manipulate the bits in the shadow copy, and then write
* the shadow copy to qlogic_ib.
*
* We deliberately make most of these 32 bits, since they have
* restricted range. For any that we read, we won't to generate 32
* bit accesses, since Opteron will generate 2 separate 32 bit HT
* transactions for a 64 bit read, and we want to avoid unnecessary
* bus transactions.
*/
/* This is the 64 bit group */
/* last ibcstatus. opaque outside chip-specific code */
u64 lastibcstat;
/* these are the "32 bit" regs */
/*
* the following two are 32-bit bitmasks, but {test,clear,set}_bit
* all expect bit fields to be "unsigned long"
*/
unsigned long p_rcvctrl; /* shadow per-port rcvctrl */
unsigned long p_sendctrl; /* shadow per-port sendctrl */
u32 ibmtu; /* The MTU programmed for this unit */
/*
* Current max size IB packet (in bytes) including IB headers, that
* we can send. Changes when ibmtu changes.
*/
u32 ibmaxlen;
/*
* ibmaxlen at init time, limited by chip and by receive buffer
* size. Not changed after init.
*/
u32 init_ibmaxlen;
/* LID programmed for this instance */
u16 lid;
/* list of pkeys programmed; 0 if not set */
u16 pkeys[4];
/* LID mask control */
u8 lmc;
u8 link_width_supported;
u8 link_speed_supported;
u8 link_width_enabled;
u8 link_speed_enabled;
u8 link_width_active;
u8 link_speed_active;
u8 vls_supported;
u8 vls_operational;
/* Rx Polarity inversion (compensate for ~tx on partner) */
u8 rx_pol_inv;
u8 hw_pidx; /* physical port index */
u8 port; /* IB port number and index into dd->pports - 1 */
u8 delay_mult;
/* used to override LED behavior */
u8 led_override; /* Substituted for normal value, if non-zero */
u16 led_override_timeoff; /* delta to next timer event */
u8 led_override_vals[2]; /* Alternates per blink-frame */
u8 led_override_phase; /* Just counts, LSB picks from vals[] */
atomic_t led_override_timer_active;
/* Used to flash LEDs in override mode */
struct timer_list led_override_timer;
struct xmit_wait cong_stats;
struct timer_list symerr_clear_timer;
/* Synchronize access between driver writes and sysfs reads */
spinlock_t cc_shadow_lock
____cacheline_aligned_in_smp;
/* Shadow copy of the congestion control table */
struct cc_table_shadow *ccti_entries_shadow;
/* Shadow copy of the congestion control entries */
struct ib_cc_congestion_setting_attr_shadow *congestion_entries_shadow;
/* List of congestion control table entries */
struct ib_cc_table_entry_shadow *ccti_entries;
/* 16 congestion entries with each entry corresponding to a SL */
struct ib_cc_congestion_entry_shadow *congestion_entries;
/* Maximum number of congestion control entries that the agent expects
* the manager to send.
*/
u16 cc_supported_table_entries;
/* Total number of congestion control table entries */
u16 total_cct_entry;
/* Bit map identifying service level */
u16 cc_sl_control_map;
/* maximum congestion control table index */
u16 ccti_limit;
/* CA's max number of 64 entry units in the congestion control table */
u8 cc_max_table_entries;
};
/* Observers. Not to be taken lightly, possibly not to ship. */
/*
* If a diag read or write is to (bottom <= offset <= top),
* the "hoook" is called, allowing, e.g. shadows to be
* updated in sync with the driver. struct diag_observer
* is the "visible" part.
*/
struct diag_observer;
typedef int (*diag_hook) (struct qib_devdata *dd,
const struct diag_observer *op,
u32 offs, u64 *data, u64 mask, int only_32);
struct diag_observer {
diag_hook hook;
u32 bottom;
u32 top;
};
extern int qib_register_observer(struct qib_devdata *dd,
const struct diag_observer *op);
/* Only declared here, not defined. Private to diags */
struct diag_observer_list_elt;
/* device data struct now contains only "general per-device" info.
* fields related to a physical IB port are in a qib_pportdata struct,
* described above) while fields only used by a particular chip-type are in
* a qib_chipdata struct, whose contents are opaque to this file.
*/
struct qib_devdata {
struct qib_ibdev verbs_dev; /* must be first */
struct list_head list;
/* pointers to related structs for this device */
/* pci access data structure */
struct pci_dev *pcidev;
struct cdev *user_cdev;
struct cdev *diag_cdev;
struct device *user_device;
struct device *diag_device;
/* mem-mapped pointer to base of chip regs */
u64 __iomem *kregbase;
/* end of mem-mapped chip space excluding sendbuf and user regs */
u64 __iomem *kregend;
/* physical address of chip for io_remap, etc. */
resource_size_t physaddr;
/* qib_cfgctxts pointers */
struct qib_ctxtdata **rcd; /* Receive Context Data */
/* qib_pportdata, points to array of (physical) port-specific
* data structs, indexed by pidx (0..n-1)
*/
struct qib_pportdata *pport;
struct qib_chip_specific *cspec; /* chip-specific */
/* kvirt address of 1st 2k pio buffer */
void __iomem *pio2kbase;
/* kvirt address of 1st 4k pio buffer */
void __iomem *pio4kbase;
/* mem-mapped pointer to base of PIO buffers (if using WC PAT) */
void __iomem *piobase;
/* mem-mapped pointer to base of user chip regs (if using WC PAT) */
u64 __iomem *userbase;
void __iomem *piovl15base; /* base of VL15 buffers, if not WC */
/*
* points to area where PIOavail registers will be DMA'ed.
* Has to be on a page of it's own, because the page will be
* mapped into user program space. This copy is *ONLY* ever
* written by DMA, not by the driver! Need a copy per device
* when we get to multiple devices
*/
volatile __le64 *pioavailregs_dma; /* DMA'ed by chip */
/* physical address where updates occur */
dma_addr_t pioavailregs_phys;
/* device-specific implementations of functions needed by
* common code. Contrary to previous consensus, we can't
* really just point to a device-specific table, because we
* may need to "bend", e.g. *_f_put_tid
*/
/* fallback to alternate interrupt type if possible */
int (*f_intr_fallback)(struct qib_devdata *);
/* hard reset chip */
int (*f_reset)(struct qib_devdata *);
void (*f_quiet_serdes)(struct qib_pportdata *);
int (*f_bringup_serdes)(struct qib_pportdata *);
int (*f_early_init)(struct qib_devdata *);
void (*f_clear_tids)(struct qib_devdata *, struct qib_ctxtdata *);
void (*f_put_tid)(struct qib_devdata *, u64 __iomem*,
u32, unsigned long);
void (*f_cleanup)(struct qib_devdata *);
void (*f_setextled)(struct qib_pportdata *, u32);
/* fill out chip-specific fields */
int (*f_get_base_info)(struct qib_ctxtdata *, struct qib_base_info *);
/* free irq */
void (*f_free_irq)(struct qib_devdata *);
struct qib_message_header *(*f_get_msgheader)
(struct qib_devdata *, __le32 *);
void (*f_config_ctxts)(struct qib_devdata *);
int (*f_get_ib_cfg)(struct qib_pportdata *, int);
int (*f_set_ib_cfg)(struct qib_pportdata *, int, u32);
int (*f_set_ib_loopback)(struct qib_pportdata *, const char *);
int (*f_get_ib_table)(struct qib_pportdata *, int, void *);
int (*f_set_ib_table)(struct qib_pportdata *, int, void *);
u32 (*f_iblink_state)(u64);
u8 (*f_ibphys_portstate)(u64);
void (*f_xgxs_reset)(struct qib_pportdata *);
/* per chip actions needed for IB Link up/down changes */
int (*f_ib_updown)(struct qib_pportdata *, int, u64);
u32 __iomem *(*f_getsendbuf)(struct qib_pportdata *, u64, u32 *);
/* Read/modify/write of GPIO pins (potentially chip-specific */
int (*f_gpio_mod)(struct qib_devdata *dd, u32 out, u32 dir,
u32 mask);
/* Enable writes to config EEPROM (if supported) */
int (*f_eeprom_wen)(struct qib_devdata *dd, int wen);
/*
* modify rcvctrl shadow[s] and write to appropriate chip-regs.
* see above QIB_RCVCTRL_xxx_ENB/DIS for operations.
* (ctxt == -1) means "all contexts", only meaningful for
* clearing. Could remove if chip_spec shutdown properly done.
*/
void (*f_rcvctrl)(struct qib_pportdata *, unsigned int op,
int ctxt);
/* Read/modify/write sendctrl appropriately for op and port. */
void (*f_sendctrl)(struct qib_pportdata *, u32 op);
void (*f_set_intr_state)(struct qib_devdata *, u32);
void (*f_set_armlaunch)(struct qib_devdata *, u32);
void (*f_wantpiobuf_intr)(struct qib_devdata *, u32);
int (*f_late_initreg)(struct qib_devdata *);
int (*f_init_sdma_regs)(struct qib_pportdata *);
u16 (*f_sdma_gethead)(struct qib_pportdata *);
int (*f_sdma_busy)(struct qib_pportdata *);
void (*f_sdma_update_tail)(struct qib_pportdata *, u16);
void (*f_sdma_set_desc_cnt)(struct qib_pportdata *, unsigned);
void (*f_sdma_sendctrl)(struct qib_pportdata *, unsigned);
void (*f_sdma_hw_clean_up)(struct qib_pportdata *);
void (*f_sdma_hw_start_up)(struct qib_pportdata *);
void (*f_sdma_init_early)(struct qib_pportdata *);
void (*f_set_cntr_sample)(struct qib_pportdata *, u32, u32);
void (*f_update_usrhead)(struct qib_ctxtdata *, u64, u32, u32, u32);
u32 (*f_hdrqempty)(struct qib_ctxtdata *);
u64 (*f_portcntr)(struct qib_pportdata *, u32);
u32 (*f_read_cntrs)(struct qib_devdata *, loff_t, char **,
u64 **);
u32 (*f_read_portcntrs)(struct qib_devdata *, loff_t, u32,
char **, u64 **);
u32 (*f_setpbc_control)(struct qib_pportdata *, u32, u8, u8);
void (*f_initvl15_bufs)(struct qib_devdata *);
void (*f_init_ctxt)(struct qib_ctxtdata *);
void (*f_txchk_change)(struct qib_devdata *, u32, u32, u32,
struct qib_ctxtdata *);
void (*f_writescratch)(struct qib_devdata *, u32);
int (*f_tempsense_rd)(struct qib_devdata *, int regnum);
#ifdef CONFIG_INFINIBAND_QIB_DCA
int (*f_notify_dca)(struct qib_devdata *, unsigned long event);
#endif
char *boardname; /* human readable board info */
/* template for writing TIDs */
u64 tidtemplate;
/* value to write to free TIDs */
u64 tidinvalid;
/* number of registers used for pioavail */
u32 pioavregs;
/* device (not port) flags, basically device capabilities */
u32 flags;
/* last buffer for user use */
u32 lastctxt_piobuf;
/* saturating counter of (non-port-specific) device interrupts */
u32 int_counter;
/* pio bufs allocated per ctxt */
u32 pbufsctxt;
/* if remainder on bufs/ctxt, ctxts < extrabuf get 1 extra */
u32 ctxts_extrabuf;
/*
* number of ctxts configured as max; zero is set to number chip
* supports, less gives more pio bufs/ctxt, etc.
*/
u32 cfgctxts;
/*
* number of ctxts available for PSM open
*/
u32 freectxts;
/*
* hint that we should update pioavailshadow before
* looking for a PIO buffer
*/
u32 upd_pio_shadow;
/* internal debugging stats */
u32 maxpkts_call;
u32 avgpkts_call;
u64 nopiobufs;
/* PCI Vendor ID (here for NodeInfo) */
u16 vendorid;
/* PCI Device ID (here for NodeInfo) */
u16 deviceid;
/* for write combining settings */
unsigned long wc_cookie;
unsigned long wc_base;
unsigned long wc_len;
/* shadow copy of struct page *'s for exp tid pages */
struct page **pageshadow;
/* shadow copy of dma handles for exp tid pages */
dma_addr_t *physshadow;
u64 __iomem *egrtidbase;
spinlock_t sendctrl_lock; /* protect changes to sendctrl shadow */
/* around rcd and (user ctxts) ctxt_cnt use (intr vs free) */
spinlock_t uctxt_lock; /* rcd and user context changes */
/*
* per unit status, see also portdata statusp
* mapped readonly into user processes so they can get unit and
* IB link status cheaply
*/
u64 *devstatusp;
char *freezemsg; /* freeze msg if hw error put chip in freeze */
u32 freezelen; /* max length of freezemsg */
/* timer used to prevent stats overflow, error throttling, etc. */
struct timer_list stats_timer;
/* timer to verify interrupts work, and fallback if possible */
struct timer_list intrchk_timer;
unsigned long ureg_align; /* user register alignment */
/*
* Protects pioavailshadow, pioavailkernel, pio_need_disarm, and
* pio_writing.
*/
spinlock_t pioavail_lock;
/*
* index of last buffer to optimize search for next
*/
u32 last_pio;
/*
* min kernel pio buffer to optimize search
*/
u32 min_kernel_pio;
/*
* Shadow copies of registers; size indicates read access size.
* Most of them are readonly, but some are write-only register,
* where we manipulate the bits in the shadow copy, and then write
* the shadow copy to qlogic_ib.
*
* We deliberately make most of these 32 bits, since they have
* restricted range. For any that we read, we won't to generate 32
* bit accesses, since Opteron will generate 2 separate 32 bit HT
* transactions for a 64 bit read, and we want to avoid unnecessary
* bus transactions.
*/
/* This is the 64 bit group */
unsigned long pioavailshadow[6];
/* bitmap of send buffers available for the kernel to use with PIO. */
unsigned long pioavailkernel[6];
/* bitmap of send buffers which need to be disarmed. */
unsigned long pio_need_disarm[3];
/* bitmap of send buffers which are being written to. */
unsigned long pio_writing[3];
/* kr_revision shadow */
u64 revision;
/* Base GUID for device (from eeprom, network order) */
__be64 base_guid;
/*
* kr_sendpiobufbase value (chip offset of pio buffers), and the
* base of the 2KB buffer s(user processes only use 2K)
*/
u64 piobufbase;
u32 pio2k_bufbase;
/* these are the "32 bit" regs */
/* number of GUIDs in the flash for this interface */
u32 nguid;
/*
* the following two are 32-bit bitmasks, but {test,clear,set}_bit
* all expect bit fields to be "unsigned long"
*/
unsigned long rcvctrl; /* shadow per device rcvctrl */
unsigned long sendctrl; /* shadow per device sendctrl */
/* value we put in kr_rcvhdrcnt */
u32 rcvhdrcnt;
/* value we put in kr_rcvhdrsize */
u32 rcvhdrsize;
/* value we put in kr_rcvhdrentsize */
u32 rcvhdrentsize;
/* kr_ctxtcnt value */
u32 ctxtcnt;
/* kr_pagealign value */
u32 palign;
/* number of "2KB" PIO buffers */
u32 piobcnt2k;
/* size in bytes of "2KB" PIO buffers */
u32 piosize2k;
/* max usable size in dwords of a "2KB" PIO buffer before going "4KB" */
u32 piosize2kmax_dwords;
/* number of "4KB" PIO buffers */
u32 piobcnt4k;
/* size in bytes of "4KB" PIO buffers */
u32 piosize4k;
/* kr_rcvegrbase value */
u32 rcvegrbase;
/* kr_rcvtidbase value */
u32 rcvtidbase;
/* kr_rcvtidcnt value */
u32 rcvtidcnt;
/* kr_userregbase */
u32 uregbase;
/* shadow the control register contents */
u32 control;
/* chip address space used by 4k pio buffers */
u32 align4k;
/* size of each rcvegrbuffer */
u16 rcvegrbufsize;
/* log2 of above */
u16 rcvegrbufsize_shift;
/* localbus width (1, 2,4,8,16,32) from config space */
u32 lbus_width;
/* localbus speed in MHz */
u32 lbus_speed;
int unit; /* unit # of this chip */
/* start of CHIP_SPEC move to chipspec, but need code changes */
/* low and high portions of MSI capability/vector */
u32 msi_lo;
/* saved after PCIe init for restore after reset */
u32 msi_hi;
/* MSI data (vector) saved for restore */
u16 msi_data;
/* so we can rewrite it after a chip reset */
u32 pcibar0;
/* so we can rewrite it after a chip reset */
u32 pcibar1;
u64 rhdrhead_intr_off;
/*
* ASCII serial number, from flash, large enough for original
* all digit strings, and longer QLogic serial number format
*/
u8 serial[16];
/* human readable board version */
u8 boardversion[96];
u8 lbus_info[32]; /* human readable localbus info */
/* chip major rev, from qib_revision */
u8 majrev;
/* chip minor rev, from qib_revision */
u8 minrev;
/* Misc small ints */
/* Number of physical ports available */
u8 num_pports;
/* Lowest context number which can be used by user processes */
u8 first_user_ctxt;
u8 n_krcv_queues;
u8 qpn_mask;
u8 skip_kctxt_mask;
u16 rhf_offset; /* offset of RHF within receive header entry */
/*
* GPIO pins for twsi-connected devices, and device code for eeprom
*/
u8 gpio_sda_num;
u8 gpio_scl_num;
u8 twsi_eeprom_dev;
u8 board_atten;
/* Support (including locks) for EEPROM logging of errors and time */
/* control access to actual counters, timer */
spinlock_t eep_st_lock;
/* control high-level access to EEPROM */
struct mutex eep_lock;
uint64_t traffic_wds;
/* active time is kept in seconds, but logged in hours */
atomic_t active_time;
/* Below are nominal shadow of EEPROM, new since last EEPROM update */
uint8_t eep_st_errs[QIB_EEP_LOG_CNT];
uint8_t eep_st_new_errs[QIB_EEP_LOG_CNT];
uint16_t eep_hrs;
/*
* masks for which bits of errs, hwerrs that cause
* each of the counters to increment.
*/
struct qib_eep_log_mask eep_st_masks[QIB_EEP_LOG_CNT];
struct qib_diag_client *diag_client;
spinlock_t qib_diag_trans_lock; /* protect diag observer ops */
struct diag_observer_list_elt *diag_observer_list;
u8 psxmitwait_supported;
/* cycle length of PS* counters in HW (in picoseconds) */
u16 psxmitwait_check_rate;
/* high volume overflow errors defered to tasklet */
struct tasklet_struct error_tasklet;
/* per device cq worker */
struct kthread_worker *worker;
int assigned_node_id; /* NUMA node closest to HCA */
};
/* hol_state values */
#define QIB_HOL_UP 0
#define QIB_HOL_INIT 1
#define QIB_SDMA_SENDCTRL_OP_ENABLE (1U << 0)
#define QIB_SDMA_SENDCTRL_OP_INTENABLE (1U << 1)
#define QIB_SDMA_SENDCTRL_OP_HALT (1U << 2)
#define QIB_SDMA_SENDCTRL_OP_CLEANUP (1U << 3)
#define QIB_SDMA_SENDCTRL_OP_DRAIN (1U << 4)
/* operation types for f_txchk_change() */
#define TXCHK_CHG_TYPE_DIS1 3
#define TXCHK_CHG_TYPE_ENAB1 2
#define TXCHK_CHG_TYPE_KERN 1
#define TXCHK_CHG_TYPE_USER 0
#define QIB_CHASE_TIME msecs_to_jiffies(145)
#define QIB_CHASE_DIS_TIME msecs_to_jiffies(160)
/* Private data for file operations */
struct qib_filedata {
struct qib_ctxtdata *rcd;
unsigned subctxt;
unsigned tidcursor;
struct qib_user_sdma_queue *pq;
int rec_cpu_num; /* for cpu affinity; -1 if none */
};
extern struct list_head qib_dev_list;
extern spinlock_t qib_devs_lock;
extern struct qib_devdata *qib_lookup(int unit);
extern u32 qib_cpulist_count;
extern unsigned long *qib_cpulist;
extern unsigned qib_wc_pat;
extern unsigned qib_cc_table_size;
int qib_init(struct qib_devdata *, int);
int init_chip_wc_pat(struct qib_devdata *dd, u32);
int qib_enable_wc(struct qib_devdata *dd);
void qib_disable_wc(struct qib_devdata *dd);
int qib_count_units(int *npresentp, int *nupp);
int qib_count_active_units(void);
int qib_cdev_init(int minor, const char *name,
const struct file_operations *fops,
struct cdev **cdevp, struct device **devp);
void qib_cdev_cleanup(struct cdev **cdevp, struct device **devp);
int qib_dev_init(void);
void qib_dev_cleanup(void);
int qib_diag_add(struct qib_devdata *);
void qib_diag_remove(struct qib_devdata *);
void qib_handle_e_ibstatuschanged(struct qib_pportdata *, u64);
void qib_sdma_update_tail(struct qib_pportdata *, u16); /* hold sdma_lock */
int qib_decode_err(struct qib_devdata *dd, char *buf, size_t blen, u64 err);
void qib_bad_intrstatus(struct qib_devdata *);
void qib_handle_urcv(struct qib_devdata *, u64);
/* clean up any per-chip chip-specific stuff */
void qib_chip_cleanup(struct qib_devdata *);
/* clean up any chip type-specific stuff */
void qib_chip_done(void);
/* check to see if we have to force ordering for write combining */
int qib_unordered_wc(void);
void qib_pio_copy(void __iomem *to, const void *from, size_t count);
void qib_disarm_piobufs(struct qib_devdata *, unsigned, unsigned);
int qib_disarm_piobufs_ifneeded(struct qib_ctxtdata *);
void qib_disarm_piobufs_set(struct qib_devdata *, unsigned long *, unsigned);
void qib_cancel_sends(struct qib_pportdata *);
int qib_create_rcvhdrq(struct qib_devdata *, struct qib_ctxtdata *);
int qib_setup_eagerbufs(struct qib_ctxtdata *);
void qib_set_ctxtcnt(struct qib_devdata *);
int qib_create_ctxts(struct qib_devdata *dd);
struct qib_ctxtdata *qib_create_ctxtdata(struct qib_pportdata *, u32, int);
void qib_init_pportdata(struct qib_pportdata *, struct qib_devdata *, u8, u8);
void qib_free_ctxtdata(struct qib_devdata *, struct qib_ctxtdata *);
u32 qib_kreceive(struct qib_ctxtdata *, u32 *, u32 *);
int qib_reset_device(int);
int qib_wait_linkstate(struct qib_pportdata *, u32, int);
int qib_set_linkstate(struct qib_pportdata *, u8);
int qib_set_mtu(struct qib_pportdata *, u16);
int qib_set_lid(struct qib_pportdata *, u32, u8);
void qib_hol_down(struct qib_pportdata *);
void qib_hol_init(struct qib_pportdata *);
void qib_hol_up(struct qib_pportdata *);
void qib_hol_event(unsigned long);
void qib_disable_after_error(struct qib_devdata *);
int qib_set_uevent_bits(struct qib_pportdata *, const int);
/* for use in system calls, where we want to know device type, etc. */
#define ctxt_fp(fp) \
(((struct qib_filedata *)(fp)->private_data)->rcd)
#define subctxt_fp(fp) \
(((struct qib_filedata *)(fp)->private_data)->subctxt)
#define tidcursor_fp(fp) \
(((struct qib_filedata *)(fp)->private_data)->tidcursor)
#define user_sdma_queue_fp(fp) \
(((struct qib_filedata *)(fp)->private_data)->pq)
static inline struct qib_devdata *dd_from_ppd(struct qib_pportdata *ppd)
{
return ppd->dd;
}
static inline struct qib_devdata *dd_from_dev(struct qib_ibdev *dev)
{
return container_of(dev, struct qib_devdata, verbs_dev);
}
static inline struct qib_devdata *dd_from_ibdev(struct ib_device *ibdev)
{
return dd_from_dev(to_idev(ibdev));
}
static inline struct qib_pportdata *ppd_from_ibp(struct qib_ibport *ibp)
{
return container_of(ibp, struct qib_pportdata, ibport_data);
}
static inline struct qib_ibport *to_iport(struct ib_device *ibdev, u8 port)
{
struct qib_devdata *dd = dd_from_ibdev(ibdev);
unsigned pidx = port - 1; /* IB number port from 1, hdw from 0 */
WARN_ON(pidx >= dd->num_pports);
return &dd->pport[pidx].ibport_data;
}
/*
* values for dd->flags (_device_ related flags) and
*/
#define QIB_HAS_LINK_LATENCY 0x1 /* supports link latency (IB 1.2) */
#define QIB_INITTED 0x2 /* chip and driver up and initted */
#define QIB_DOING_RESET 0x4 /* in the middle of doing chip reset */
#define QIB_PRESENT 0x8 /* chip accesses can be done */
#define QIB_PIO_FLUSH_WC 0x10 /* Needs Write combining flush for PIO */
#define QIB_HAS_THRESH_UPDATE 0x40
#define QIB_HAS_SDMA_TIMEOUT 0x80
#define QIB_USE_SPCL_TRIG 0x100 /* SpecialTrigger launch enabled */
#define QIB_NODMA_RTAIL 0x200 /* rcvhdrtail register DMA enabled */
#define QIB_HAS_INTX 0x800 /* Supports INTx interrupts */
#define QIB_HAS_SEND_DMA 0x1000 /* Supports Send DMA */
#define QIB_HAS_VLSUPP 0x2000 /* Supports multiple VLs; PBC different */
#define QIB_HAS_HDRSUPP 0x4000 /* Supports header suppression */
#define QIB_BADINTR 0x8000 /* severe interrupt problems */
#define QIB_DCA_ENABLED 0x10000 /* Direct Cache Access enabled */
#define QIB_HAS_QSFP 0x20000 /* device (card instance) has QSFP */
/*
* values for ppd->lflags (_ib_port_ related flags)
*/
#define QIBL_LINKV 0x1 /* IB link state valid */
#define QIBL_LINKDOWN 0x8 /* IB link is down */
#define QIBL_LINKINIT 0x10 /* IB link level is up */
#define QIBL_LINKARMED 0x20 /* IB link is ARMED */
#define QIBL_LINKACTIVE 0x40 /* IB link is ACTIVE */
/* leave a gap for more IB-link state */
#define QIBL_IB_AUTONEG_INPROG 0x1000 /* non-IBTA DDR/QDR neg active */
#define QIBL_IB_AUTONEG_FAILED 0x2000 /* non-IBTA DDR/QDR neg failed */
#define QIBL_IB_LINK_DISABLED 0x4000 /* Linkdown-disable forced,
* Do not try to bring up */
#define QIBL_IB_FORCE_NOTIFY 0x8000 /* force notify on next ib change */
/* IB dword length mask in PBC (lower 11 bits); same for all chips */
#define QIB_PBC_LENGTH_MASK ((1 << 11) - 1)
/* ctxt_flag bit offsets */
/* waiting for a packet to arrive */
#define QIB_CTXT_WAITING_RCV 2
/* master has not finished initializing */
#define QIB_CTXT_MASTER_UNINIT 4
/* waiting for an urgent packet to arrive */
#define QIB_CTXT_WAITING_URG 5
/* free up any allocated data at closes */
void qib_free_data(struct qib_ctxtdata *dd);
void qib_chg_pioavailkernel(struct qib_devdata *, unsigned, unsigned,
u32, struct qib_ctxtdata *);
struct qib_devdata *qib_init_iba7322_funcs(struct pci_dev *,
const struct pci_device_id *);
struct qib_devdata *qib_init_iba7220_funcs(struct pci_dev *,
const struct pci_device_id *);
struct qib_devdata *qib_init_iba6120_funcs(struct pci_dev *,
const struct pci_device_id *);
void qib_free_devdata(struct qib_devdata *);
struct qib_devdata *qib_alloc_devdata(struct pci_dev *pdev, size_t extra);
#define QIB_TWSI_NO_DEV 0xFF
/* Below qib_twsi_ functions must be called with eep_lock held */
int qib_twsi_reset(struct qib_devdata *dd);
int qib_twsi_blk_rd(struct qib_devdata *dd, int dev, int addr, void *buffer,
int len);
int qib_twsi_blk_wr(struct qib_devdata *dd, int dev, int addr,
const void *buffer, int len);
void qib_get_eeprom_info(struct qib_devdata *);
int qib_update_eeprom_log(struct qib_devdata *dd);
void qib_inc_eeprom_err(struct qib_devdata *dd, u32 eidx, u32 incr);
void qib_dump_lookup_output_queue(struct qib_devdata *);
void qib_force_pio_avail_update(struct qib_devdata *);
void qib_clear_symerror_on_linkup(unsigned long opaque);
/*
* Set LED override, only the two LSBs have "public" meaning, but
* any non-zero value substitutes them for the Link and LinkTrain
* LED states.
*/
#define QIB_LED_PHYS 1 /* Physical (linktraining) GREEN LED */
#define QIB_LED_LOG 2 /* Logical (link) YELLOW LED */
void qib_set_led_override(struct qib_pportdata *ppd, unsigned int val);
/* send dma routines */
int qib_setup_sdma(struct qib_pportdata *);
void qib_teardown_sdma(struct qib_pportdata *);
void __qib_sdma_intr(struct qib_pportdata *);
void qib_sdma_intr(struct qib_pportdata *);
int qib_sdma_verbs_send(struct qib_pportdata *, struct qib_sge_state *,
u32, struct qib_verbs_txreq *);
/* ppd->sdma_lock should be locked before calling this. */
int qib_sdma_make_progress(struct qib_pportdata *dd);
static inline int qib_sdma_empty(const struct qib_pportdata *ppd)
{
return ppd->sdma_descq_added == ppd->sdma_descq_removed;
}
/* must be called under qib_sdma_lock */
static inline u16 qib_sdma_descq_freecnt(const struct qib_pportdata *ppd)
{
return ppd->sdma_descq_cnt -
(ppd->sdma_descq_added - ppd->sdma_descq_removed) - 1;
}
static inline int __qib_sdma_running(struct qib_pportdata *ppd)
{
return ppd->sdma_state.current_state == qib_sdma_state_s99_running;
}
int qib_sdma_running(struct qib_pportdata *);
void __qib_sdma_process_event(struct qib_pportdata *, enum qib_sdma_events);
void qib_sdma_process_event(struct qib_pportdata *, enum qib_sdma_events);
/*
* number of words used for protocol header if not set by qib_userinit();
*/
#define QIB_DFLT_RCVHDRSIZE 9
/*
* We need to be able to handle an IB header of at least 24 dwords.
* We need the rcvhdrq large enough to handle largest IB header, but
* still have room for a 2KB MTU standard IB packet.
* Additionally, some processor/memory controller combinations
* benefit quite strongly from having the DMA'ed data be cacheline
* aligned and a cacheline multiple, so we set the size to 32 dwords
* (2 64-byte primary cachelines for pretty much all processors of
* interest). The alignment hurts nothing, other than using somewhat
* more memory.
*/
#define QIB_RCVHDR_ENTSIZE 32
int qib_get_user_pages(unsigned long, size_t, struct page **);
void qib_release_user_pages(struct page **, size_t);
int qib_eeprom_read(struct qib_devdata *, u8, void *, int);
int qib_eeprom_write(struct qib_devdata *, u8, const void *, int);
u32 __iomem *qib_getsendbuf_range(struct qib_devdata *, u32 *, u32, u32);
void qib_sendbuf_done(struct qib_devdata *, unsigned);
static inline void qib_clear_rcvhdrtail(const struct qib_ctxtdata *rcd)
{
*((u64 *) rcd->rcvhdrtail_kvaddr) = 0ULL;
}
static inline u32 qib_get_rcvhdrtail(const struct qib_ctxtdata *rcd)
{
/*
* volatile because it's a DMA target from the chip, routine is
* inlined, and don't want register caching or reordering.
*/
return (u32) le64_to_cpu(
*((volatile __le64 *)rcd->rcvhdrtail_kvaddr)); /* DMA'ed */
}
static inline u32 qib_get_hdrqtail(const struct qib_ctxtdata *rcd)
{
const struct qib_devdata *dd = rcd->dd;
u32 hdrqtail;
if (dd->flags & QIB_NODMA_RTAIL) {
__le32 *rhf_addr;
u32 seq;
rhf_addr = (__le32 *) rcd->rcvhdrq +
rcd->head + dd->rhf_offset;
seq = qib_hdrget_seq(rhf_addr);
hdrqtail = rcd->head;
if (seq == rcd->seq_cnt)
hdrqtail++;
} else
hdrqtail = qib_get_rcvhdrtail(rcd);
return hdrqtail;
}
/*
* sysfs interface.
*/
extern const char ib_qib_version[];
int qib_device_create(struct qib_devdata *);
void qib_device_remove(struct qib_devdata *);
int qib_create_port_files(struct ib_device *ibdev, u8 port_num,
struct kobject *kobj);
int qib_verbs_register_sysfs(struct qib_devdata *);
void qib_verbs_unregister_sysfs(struct qib_devdata *);
/* Hook for sysfs read of QSFP */
extern int qib_qsfp_dump(struct qib_pportdata *ppd, char *buf, int len);
int __init qib_init_qibfs(void);
int __exit qib_exit_qibfs(void);
int qibfs_add(struct qib_devdata *);
int qibfs_remove(struct qib_devdata *);
int qib_pcie_init(struct pci_dev *, const struct pci_device_id *);
int qib_pcie_ddinit(struct qib_devdata *, struct pci_dev *,
const struct pci_device_id *);
void qib_pcie_ddcleanup(struct qib_devdata *);
int qib_pcie_params(struct qib_devdata *, u32, u32 *, struct qib_msix_entry *);
int qib_reinit_intr(struct qib_devdata *);
void qib_enable_intx(struct pci_dev *);
void qib_nomsi(struct qib_devdata *);
void qib_nomsix(struct qib_devdata *);
void qib_pcie_getcmd(struct qib_devdata *, u16 *, u8 *, u8 *);
void qib_pcie_reenable(struct qib_devdata *, u16, u8, u8);
/*
* dma_addr wrappers - all 0's invalid for hw
*/
dma_addr_t qib_map_page(struct pci_dev *, struct page *, unsigned long,
size_t, int);
const char *qib_get_unit_name(int unit);
/*
* Flush write combining store buffers (if present) and perform a write
* barrier.
*/
#if defined(CONFIG_X86_64)
#define qib_flush_wc() asm volatile("sfence" : : : "memory")
#else
#define qib_flush_wc() wmb() /* no reorder around wc flush */
#endif
/* global module parameter variables */
extern unsigned qib_ibmtu;
extern ushort qib_cfgctxts;
extern ushort qib_num_cfg_vls;
extern ushort qib_mini_init; /* If set, do few (ideally 0) writes to chip */
extern unsigned qib_n_krcv_queues;
extern unsigned qib_sdma_fetch_arb;
extern unsigned qib_compat_ddr_negotiate;
extern int qib_special_trigger;
extern unsigned qib_numa_aware;
extern struct mutex qib_mutex;
/* Number of seconds before our card status check... */
#define STATUS_TIMEOUT 60
#define QIB_DRV_NAME "ib_qib"
#define QIB_USER_MINOR_BASE 0
#define QIB_TRACE_MINOR 127
#define QIB_DIAGPKT_MINOR 128
#define QIB_DIAG_MINOR_BASE 129
#define QIB_NMINORS 255
#define PCI_VENDOR_ID_PATHSCALE 0x1fc1
#define PCI_VENDOR_ID_QLOGIC 0x1077
#define PCI_DEVICE_ID_QLOGIC_IB_6120 0x10
#define PCI_DEVICE_ID_QLOGIC_IB_7220 0x7220
#define PCI_DEVICE_ID_QLOGIC_IB_7322 0x7322
/*
* qib_early_err is used (only!) to print early errors before devdata is
* allocated, or when dd->pcidev may not be valid, and at the tail end of
* cleanup when devdata may have been freed, etc. qib_dev_porterr is
* the same as qib_dev_err, but is used when the message really needs
* the IB port# to be definitive as to what's happening..
* All of these go to the trace log, and the trace log entry is done
* first to avoid possible serial port delays from printk.
*/
#define qib_early_err(dev, fmt, ...) \
dev_err(dev, fmt, ##__VA_ARGS__)
#define qib_dev_err(dd, fmt, ...) \
dev_err(&(dd)->pcidev->dev, "%s: " fmt, \
qib_get_unit_name((dd)->unit), ##__VA_ARGS__)
#define qib_dev_warn(dd, fmt, ...) \
dev_warn(&(dd)->pcidev->dev, "%s: " fmt, \
qib_get_unit_name((dd)->unit), ##__VA_ARGS__)
#define qib_dev_porterr(dd, port, fmt, ...) \
dev_err(&(dd)->pcidev->dev, "%s: IB%u:%u " fmt, \
qib_get_unit_name((dd)->unit), (dd)->unit, (port), \
##__VA_ARGS__)
#define qib_devinfo(pcidev, fmt, ...) \
dev_info(&(pcidev)->dev, fmt, ##__VA_ARGS__)
/*
* this is used for formatting hw error messages...
*/
struct qib_hwerror_msgs {
u64 mask;
const char *msg;
size_t sz;
};
#define QLOGIC_IB_HWE_MSG(a, b) { .mask = a, .msg = b }
/* in qib_intr.c... */
void qib_format_hwerrors(u64 hwerrs,
const struct qib_hwerror_msgs *hwerrmsgs,
size_t nhwerrmsgs, char *msg, size_t lmsg);
#endif /* _QIB_KERNEL_H */
|