1 /* SPDX-License-Identifier: GPL-2.0 OR Linux-O 1 /* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */ 2 /* 2 /* 3 * Copyright (c) 2004 Mellanox Technologies Lt 3 * Copyright (c) 2004 Mellanox Technologies Ltd. All rights reserved. 4 * Copyright (c) 2004 Infinicon Corporation. 4 * Copyright (c) 2004 Infinicon Corporation. All rights reserved. 5 * Copyright (c) 2004 Intel Corporation. All 5 * Copyright (c) 2004 Intel Corporation. All rights reserved. 6 * Copyright (c) 2004 Topspin Corporation. Al 6 * Copyright (c) 2004 Topspin Corporation. All rights reserved. 7 * Copyright (c) 2004-2006 Voltaire Corporatio 7 * Copyright (c) 2004-2006 Voltaire Corporation. All rights reserved. 8 */ 8 */ 9 9 10 #ifndef IB_MAD_H 10 #ifndef IB_MAD_H 11 #define IB_MAD_H 11 #define IB_MAD_H 12 12 13 #include <linux/list.h> 13 #include <linux/list.h> 14 14 15 #include <rdma/ib_verbs.h> 15 #include <rdma/ib_verbs.h> 16 #include <uapi/rdma/ib_user_mad.h> 16 #include <uapi/rdma/ib_user_mad.h> 17 17 18 /* Management base versions */ 18 /* Management base versions */ 19 #define IB_MGMT_BASE_VERSION 19 #define IB_MGMT_BASE_VERSION 1 20 #define OPA_MGMT_BASE_VERSION 20 #define OPA_MGMT_BASE_VERSION 0x80 21 21 22 #define OPA_SM_CLASS_VERSION 22 #define OPA_SM_CLASS_VERSION 0x80 23 23 24 /* Management classes */ 24 /* Management classes */ 25 #define IB_MGMT_CLASS_SUBN_LID_ROUTED 25 #define IB_MGMT_CLASS_SUBN_LID_ROUTED 0x01 26 #define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE 26 #define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE 0x81 27 #define IB_MGMT_CLASS_SUBN_ADM 27 #define IB_MGMT_CLASS_SUBN_ADM 0x03 28 #define IB_MGMT_CLASS_PERF_MGMT 28 #define IB_MGMT_CLASS_PERF_MGMT 0x04 29 #define IB_MGMT_CLASS_BM 29 #define IB_MGMT_CLASS_BM 0x05 30 #define IB_MGMT_CLASS_DEVICE_MGMT 30 #define IB_MGMT_CLASS_DEVICE_MGMT 0x06 31 #define IB_MGMT_CLASS_CM 31 #define IB_MGMT_CLASS_CM 0x07 32 #define IB_MGMT_CLASS_SNMP 32 #define IB_MGMT_CLASS_SNMP 0x08 33 #define IB_MGMT_CLASS_DEVICE_ADM 33 #define IB_MGMT_CLASS_DEVICE_ADM 0x10 34 #define IB_MGMT_CLASS_BOOT_MGMT 34 #define IB_MGMT_CLASS_BOOT_MGMT 0x11 35 #define IB_MGMT_CLASS_BIS 35 #define IB_MGMT_CLASS_BIS 0x12 36 #define IB_MGMT_CLASS_CONG_MGMT 36 #define IB_MGMT_CLASS_CONG_MGMT 0x21 37 #define IB_MGMT_CLASS_VENDOR_RANGE2_START 37 #define IB_MGMT_CLASS_VENDOR_RANGE2_START 0x30 38 #define IB_MGMT_CLASS_VENDOR_RANGE2_END 38 #define IB_MGMT_CLASS_VENDOR_RANGE2_END 0x4F 39 39 40 #define IB_OPENIB_OUI 40 #define IB_OPENIB_OUI (0x001405) 41 41 42 /* Management methods */ 42 /* Management methods */ 43 #define IB_MGMT_METHOD_GET 43 #define IB_MGMT_METHOD_GET 0x01 44 #define IB_MGMT_METHOD_SET 44 #define IB_MGMT_METHOD_SET 0x02 45 #define IB_MGMT_METHOD_GET_RESP 45 #define IB_MGMT_METHOD_GET_RESP 0x81 46 #define IB_MGMT_METHOD_SEND 46 #define IB_MGMT_METHOD_SEND 0x03 47 #define IB_MGMT_METHOD_TRAP 47 #define IB_MGMT_METHOD_TRAP 0x05 48 #define IB_MGMT_METHOD_REPORT 48 #define IB_MGMT_METHOD_REPORT 0x06 49 #define IB_MGMT_METHOD_REPORT_RESP 49 #define IB_MGMT_METHOD_REPORT_RESP 0x86 50 #define IB_MGMT_METHOD_TRAP_REPRESS 50 #define IB_MGMT_METHOD_TRAP_REPRESS 0x07 51 51 52 #define IB_MGMT_METHOD_RESP 52 #define IB_MGMT_METHOD_RESP 0x80 53 #define IB_BM_ATTR_MOD_RESP 53 #define IB_BM_ATTR_MOD_RESP cpu_to_be32(1) 54 54 55 #define IB_MGMT_MAX_METHODS 55 #define IB_MGMT_MAX_METHODS 128 56 56 57 /* MAD Status field bit masks */ 57 /* MAD Status field bit masks */ 58 #define IB_MGMT_MAD_STATUS_SUCCESS 58 #define IB_MGMT_MAD_STATUS_SUCCESS 0x0000 59 #define IB_MGMT_MAD_STATUS_BUSY 59 #define IB_MGMT_MAD_STATUS_BUSY 0x0001 60 #define IB_MGMT_MAD_STATUS_REDIRECT_REQD 60 #define IB_MGMT_MAD_STATUS_REDIRECT_REQD 0x0002 61 #define IB_MGMT_MAD_STATUS_BAD_VERSION 61 #define IB_MGMT_MAD_STATUS_BAD_VERSION 0x0004 62 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD 62 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD 0x0008 63 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ 63 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ATTRIB 0x000c 64 #define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALU 64 #define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALUE 0x001c 65 65 66 /* RMPP information */ 66 /* RMPP information */ 67 #define IB_MGMT_RMPP_VERSION 67 #define IB_MGMT_RMPP_VERSION 1 68 68 69 #define IB_MGMT_RMPP_TYPE_DATA 69 #define IB_MGMT_RMPP_TYPE_DATA 1 70 #define IB_MGMT_RMPP_TYPE_ACK 70 #define IB_MGMT_RMPP_TYPE_ACK 2 71 #define IB_MGMT_RMPP_TYPE_STOP 71 #define IB_MGMT_RMPP_TYPE_STOP 3 72 #define IB_MGMT_RMPP_TYPE_ABORT 72 #define IB_MGMT_RMPP_TYPE_ABORT 4 73 73 74 #define IB_MGMT_RMPP_FLAG_ACTIVE 74 #define IB_MGMT_RMPP_FLAG_ACTIVE 1 75 #define IB_MGMT_RMPP_FLAG_FIRST 75 #define IB_MGMT_RMPP_FLAG_FIRST (1<<1) 76 #define IB_MGMT_RMPP_FLAG_LAST 76 #define IB_MGMT_RMPP_FLAG_LAST (1<<2) 77 77 78 #define IB_MGMT_RMPP_NO_RESPTIME 78 #define IB_MGMT_RMPP_NO_RESPTIME 0x1F 79 79 80 #define IB_MGMT_RMPP_STATUS_SUCCESS 80 #define IB_MGMT_RMPP_STATUS_SUCCESS 0 81 #define IB_MGMT_RMPP_STATUS_RESX 81 #define IB_MGMT_RMPP_STATUS_RESX 1 82 #define IB_MGMT_RMPP_STATUS_ABORT_MIN 82 #define IB_MGMT_RMPP_STATUS_ABORT_MIN 118 83 #define IB_MGMT_RMPP_STATUS_T2L 83 #define IB_MGMT_RMPP_STATUS_T2L 118 84 #define IB_MGMT_RMPP_STATUS_BAD_LEN 84 #define IB_MGMT_RMPP_STATUS_BAD_LEN 119 85 #define IB_MGMT_RMPP_STATUS_BAD_SEG 85 #define IB_MGMT_RMPP_STATUS_BAD_SEG 120 86 #define IB_MGMT_RMPP_STATUS_BADT 86 #define IB_MGMT_RMPP_STATUS_BADT 121 87 #define IB_MGMT_RMPP_STATUS_W2S 87 #define IB_MGMT_RMPP_STATUS_W2S 122 88 #define IB_MGMT_RMPP_STATUS_S2B 88 #define IB_MGMT_RMPP_STATUS_S2B 123 89 #define IB_MGMT_RMPP_STATUS_BAD_STATUS 89 #define IB_MGMT_RMPP_STATUS_BAD_STATUS 124 90 #define IB_MGMT_RMPP_STATUS_UNV 90 #define IB_MGMT_RMPP_STATUS_UNV 125 91 #define IB_MGMT_RMPP_STATUS_TMR 91 #define IB_MGMT_RMPP_STATUS_TMR 126 92 #define IB_MGMT_RMPP_STATUS_UNSPEC 92 #define IB_MGMT_RMPP_STATUS_UNSPEC 127 93 #define IB_MGMT_RMPP_STATUS_ABORT_MAX 93 #define IB_MGMT_RMPP_STATUS_ABORT_MAX 127 94 94 95 #define IB_QP0 0 95 #define IB_QP0 0 96 #define IB_QP1 cpu_to_be32(1) 96 #define IB_QP1 cpu_to_be32(1) 97 #define IB_QP1_QKEY 0x80010000 97 #define IB_QP1_QKEY 0x80010000 98 #define IB_QP_SET_QKEY 0x80000000 98 #define IB_QP_SET_QKEY 0x80000000 99 99 100 #define IB_DEFAULT_PKEY_PARTIAL 0x7FFF 100 #define IB_DEFAULT_PKEY_PARTIAL 0x7FFF 101 #define IB_DEFAULT_PKEY_FULL 0xFFFF 101 #define IB_DEFAULT_PKEY_FULL 0xFFFF 102 102 103 /* 103 /* 104 * Generic trap/notice types 104 * Generic trap/notice types 105 */ 105 */ 106 #define IB_NOTICE_TYPE_FATAL 0x80 106 #define IB_NOTICE_TYPE_FATAL 0x80 107 #define IB_NOTICE_TYPE_URGENT 0x81 107 #define IB_NOTICE_TYPE_URGENT 0x81 108 #define IB_NOTICE_TYPE_SECURITY 0x82 108 #define IB_NOTICE_TYPE_SECURITY 0x82 109 #define IB_NOTICE_TYPE_SM 0x83 109 #define IB_NOTICE_TYPE_SM 0x83 110 #define IB_NOTICE_TYPE_INFO 0x84 110 #define IB_NOTICE_TYPE_INFO 0x84 111 111 112 /* 112 /* 113 * Generic trap/notice producers 113 * Generic trap/notice producers 114 */ 114 */ 115 #define IB_NOTICE_PROD_CA cpu_to 115 #define IB_NOTICE_PROD_CA cpu_to_be16(1) 116 #define IB_NOTICE_PROD_SWITCH cpu_to 116 #define IB_NOTICE_PROD_SWITCH cpu_to_be16(2) 117 #define IB_NOTICE_PROD_ROUTER cpu_to 117 #define IB_NOTICE_PROD_ROUTER cpu_to_be16(3) 118 #define IB_NOTICE_PROD_CLASS_MGR cpu_to 118 #define IB_NOTICE_PROD_CLASS_MGR cpu_to_be16(4) 119 119 120 enum { 120 enum { 121 IB_MGMT_MAD_HDR = 24, 121 IB_MGMT_MAD_HDR = 24, 122 IB_MGMT_MAD_DATA = 232, 122 IB_MGMT_MAD_DATA = 232, 123 IB_MGMT_RMPP_HDR = 36, 123 IB_MGMT_RMPP_HDR = 36, 124 IB_MGMT_RMPP_DATA = 220, 124 IB_MGMT_RMPP_DATA = 220, 125 IB_MGMT_VENDOR_HDR = 40, 125 IB_MGMT_VENDOR_HDR = 40, 126 IB_MGMT_VENDOR_DATA = 216, 126 IB_MGMT_VENDOR_DATA = 216, 127 IB_MGMT_SA_HDR = 56, 127 IB_MGMT_SA_HDR = 56, 128 IB_MGMT_SA_DATA = 200, 128 IB_MGMT_SA_DATA = 200, 129 IB_MGMT_DEVICE_HDR = 64, 129 IB_MGMT_DEVICE_HDR = 64, 130 IB_MGMT_DEVICE_DATA = 192, 130 IB_MGMT_DEVICE_DATA = 192, 131 IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + I 131 IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + IB_MGMT_MAD_DATA, 132 OPA_MGMT_MAD_DATA = 2024, 132 OPA_MGMT_MAD_DATA = 2024, 133 OPA_MGMT_RMPP_DATA = 2012, 133 OPA_MGMT_RMPP_DATA = 2012, 134 OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + 134 OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + OPA_MGMT_MAD_DATA, 135 }; 135 }; 136 136 137 struct ib_mad_hdr { 137 struct ib_mad_hdr { 138 u8 base_version; 138 u8 base_version; 139 u8 mgmt_class; 139 u8 mgmt_class; 140 u8 class_version; 140 u8 class_version; 141 u8 method; 141 u8 method; 142 __be16 status; 142 __be16 status; 143 __be16 class_specific; 143 __be16 class_specific; 144 __be64 tid; 144 __be64 tid; 145 __be16 attr_id; 145 __be16 attr_id; 146 __be16 resv; 146 __be16 resv; 147 __be32 attr_mod; 147 __be32 attr_mod; 148 }; 148 }; 149 149 150 struct ib_rmpp_hdr { 150 struct ib_rmpp_hdr { 151 u8 rmpp_version; 151 u8 rmpp_version; 152 u8 rmpp_type; 152 u8 rmpp_type; 153 u8 rmpp_rtime_flags; 153 u8 rmpp_rtime_flags; 154 u8 rmpp_status; 154 u8 rmpp_status; 155 __be32 seg_num; 155 __be32 seg_num; 156 __be32 paylen_newwin; 156 __be32 paylen_newwin; 157 }; 157 }; 158 158 159 typedef u64 __bitwise ib_sa_comp_mask; 159 typedef u64 __bitwise ib_sa_comp_mask; 160 160 161 #define IB_SA_COMP_MASK(n) ((__force ib_sa_com 161 #define IB_SA_COMP_MASK(n) ((__force ib_sa_comp_mask) cpu_to_be64(1ull << (n))) 162 162 163 /* 163 /* 164 * ib_sa_hdr and ib_sa_mad structures must be 164 * ib_sa_hdr and ib_sa_mad structures must be packed because they have 165 * 64-bit fields that are only 32-bit aligned. 165 * 64-bit fields that are only 32-bit aligned. 64-bit architectures will 166 * lay them out wrong otherwise. (And unfortu 166 * lay them out wrong otherwise. (And unfortunately they are sent on 167 * the wire so we can't change the layout) 167 * the wire so we can't change the layout) 168 */ 168 */ 169 struct ib_sa_hdr { 169 struct ib_sa_hdr { 170 __be64 sm_key; 170 __be64 sm_key; 171 __be16 attr_offset; 171 __be16 attr_offset; 172 __be16 reserved; 172 __be16 reserved; 173 ib_sa_comp_mask comp_mask; 173 ib_sa_comp_mask comp_mask; 174 } __packed; 174 } __packed; 175 175 176 struct ib_mad { 176 struct ib_mad { 177 struct ib_mad_hdr mad_hdr; 177 struct ib_mad_hdr mad_hdr; 178 u8 data[IB_MGMT_M 178 u8 data[IB_MGMT_MAD_DATA]; 179 }; 179 }; 180 180 181 struct opa_mad { 181 struct opa_mad { 182 struct ib_mad_hdr mad_hdr; 182 struct ib_mad_hdr mad_hdr; 183 u8 data[OPA_MGMT_ 183 u8 data[OPA_MGMT_MAD_DATA]; 184 }; 184 }; 185 185 186 struct ib_rmpp_mad { 186 struct ib_rmpp_mad { 187 struct ib_mad_hdr mad_hdr; 187 struct ib_mad_hdr mad_hdr; 188 struct ib_rmpp_hdr rmpp_hdr; 188 struct ib_rmpp_hdr rmpp_hdr; 189 u8 data[IB_MGMT_R 189 u8 data[IB_MGMT_RMPP_DATA]; 190 }; 190 }; 191 191 192 struct opa_rmpp_mad { 192 struct opa_rmpp_mad { 193 struct ib_mad_hdr mad_hdr; 193 struct ib_mad_hdr mad_hdr; 194 struct ib_rmpp_hdr rmpp_hdr; 194 struct ib_rmpp_hdr rmpp_hdr; 195 u8 data[OPA_MGMT_ 195 u8 data[OPA_MGMT_RMPP_DATA]; 196 }; 196 }; 197 197 198 struct ib_sa_mad { 198 struct ib_sa_mad { 199 struct ib_mad_hdr mad_hdr; 199 struct ib_mad_hdr mad_hdr; 200 struct ib_rmpp_hdr rmpp_hdr; 200 struct ib_rmpp_hdr rmpp_hdr; 201 struct ib_sa_hdr sa_hdr; 201 struct ib_sa_hdr sa_hdr; 202 u8 data[IB_MGMT_S 202 u8 data[IB_MGMT_SA_DATA]; 203 } __packed; 203 } __packed; 204 204 205 struct ib_vendor_mad { 205 struct ib_vendor_mad { 206 struct ib_mad_hdr mad_hdr; 206 struct ib_mad_hdr mad_hdr; 207 struct ib_rmpp_hdr rmpp_hdr; 207 struct ib_rmpp_hdr rmpp_hdr; 208 u8 reserved; 208 u8 reserved; 209 u8 oui[3]; 209 u8 oui[3]; 210 u8 data[IB_MGMT_V 210 u8 data[IB_MGMT_VENDOR_DATA]; 211 }; 211 }; 212 212 213 #define IB_MGMT_CLASSPORTINFO_ATTR_ID cpu_to 213 #define IB_MGMT_CLASSPORTINFO_ATTR_ID cpu_to_be16(0x0001) 214 214 215 #define IB_CLASS_PORT_INFO_RESP_TIME_MASK 215 #define IB_CLASS_PORT_INFO_RESP_TIME_MASK 0x1F 216 #define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZ 216 #define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE 5 217 217 218 struct ib_class_port_info { 218 struct ib_class_port_info { 219 u8 base_version; 219 u8 base_version; 220 u8 class_version; 220 u8 class_version; 221 __be16 capability_mas 221 __be16 capability_mask; 222 /* 27 bits for cap_mask2, 5 bits for 222 /* 27 bits for cap_mask2, 5 bits for resp_time */ 223 __be32 cap_mask2_resp 223 __be32 cap_mask2_resp_time; 224 u8 redirect_gid[1 224 u8 redirect_gid[16]; 225 __be32 redirect_tcslf 225 __be32 redirect_tcslfl; 226 __be16 redirect_lid; 226 __be16 redirect_lid; 227 __be16 redirect_pkey; 227 __be16 redirect_pkey; 228 __be32 redirect_qp; 228 __be32 redirect_qp; 229 __be32 redirect_qkey; 229 __be32 redirect_qkey; 230 u8 trap_gid[16]; 230 u8 trap_gid[16]; 231 __be32 trap_tcslfl; 231 __be32 trap_tcslfl; 232 __be16 trap_lid; 232 __be16 trap_lid; 233 __be16 trap_pkey; 233 __be16 trap_pkey; 234 __be32 trap_hlqp; 234 __be32 trap_hlqp; 235 __be32 trap_qkey; 235 __be32 trap_qkey; 236 }; 236 }; 237 237 238 /* PortInfo CapabilityMask */ 238 /* PortInfo CapabilityMask */ 239 enum ib_port_capability_mask_bits { 239 enum ib_port_capability_mask_bits { 240 IB_PORT_SM = 1 << 1, 240 IB_PORT_SM = 1 << 1, 241 IB_PORT_NOTICE_SUP = 1 << 2, 241 IB_PORT_NOTICE_SUP = 1 << 2, 242 IB_PORT_TRAP_SUP = 1 << 3, 242 IB_PORT_TRAP_SUP = 1 << 3, 243 IB_PORT_OPT_IPD_SUP = 1 << 4, 243 IB_PORT_OPT_IPD_SUP = 1 << 4, 244 IB_PORT_AUTO_MIGR_SUP = 1 << 5, 244 IB_PORT_AUTO_MIGR_SUP = 1 << 5, 245 IB_PORT_SL_MAP_SUP = 1 << 6, 245 IB_PORT_SL_MAP_SUP = 1 << 6, 246 IB_PORT_MKEY_NVRAM = 1 << 7, 246 IB_PORT_MKEY_NVRAM = 1 << 7, 247 IB_PORT_PKEY_NVRAM = 1 << 8, 247 IB_PORT_PKEY_NVRAM = 1 << 8, 248 IB_PORT_LED_INFO_SUP = 1 << 9, 248 IB_PORT_LED_INFO_SUP = 1 << 9, 249 IB_PORT_SM_DISABLED = 1 << 10, 249 IB_PORT_SM_DISABLED = 1 << 10, 250 IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11, 250 IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11, 251 IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 251 IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 << 12, 252 IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14, 252 IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14, 253 IB_PORT_CAP_MASK2_SUP = 1 << 15, 253 IB_PORT_CAP_MASK2_SUP = 1 << 15, 254 IB_PORT_CM_SUP = 1 << 16, 254 IB_PORT_CM_SUP = 1 << 16, 255 IB_PORT_SNMP_TUNNEL_SUP = 1 << 17, 255 IB_PORT_SNMP_TUNNEL_SUP = 1 << 17, 256 IB_PORT_REINIT_SUP = 1 << 18, 256 IB_PORT_REINIT_SUP = 1 << 18, 257 IB_PORT_DEVICE_MGMT_SUP = 1 << 19, 257 IB_PORT_DEVICE_MGMT_SUP = 1 << 19, 258 IB_PORT_VENDOR_CLASS_SUP = 1 << 20, 258 IB_PORT_VENDOR_CLASS_SUP = 1 << 20, 259 IB_PORT_DR_NOTICE_SUP = 1 << 21, 259 IB_PORT_DR_NOTICE_SUP = 1 << 21, 260 IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22, 260 IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22, 261 IB_PORT_BOOT_MGMT_SUP = 1 << 23, 261 IB_PORT_BOOT_MGMT_SUP = 1 << 23, 262 IB_PORT_LINK_LATENCY_SUP = 1 << 24, 262 IB_PORT_LINK_LATENCY_SUP = 1 << 24, 263 IB_PORT_CLIENT_REG_SUP = 1 << 25, 263 IB_PORT_CLIENT_REG_SUP = 1 << 25, 264 IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 264 IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 26, 265 IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 265 IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 << 27, 266 IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP 266 IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP = 1 << 28, 267 IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SU 267 IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SUP = 1 << 29, 268 IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30, 268 IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30, 269 IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 3 269 IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 31, 270 }; 270 }; 271 271 272 enum ib_port_capability_mask2_bits { 272 enum ib_port_capability_mask2_bits { 273 IB_PORT_SET_NODE_DESC_SUP 273 IB_PORT_SET_NODE_DESC_SUP = 1 << 0, 274 IB_PORT_EX_PORT_INFO_EX_SUP 274 IB_PORT_EX_PORT_INFO_EX_SUP = 1 << 1, 275 IB_PORT_VIRT_SUP 275 IB_PORT_VIRT_SUP = 1 << 2, 276 IB_PORT_SWITCH_PORT_STATE_TABLE_SUP 276 IB_PORT_SWITCH_PORT_STATE_TABLE_SUP = 1 << 3, 277 IB_PORT_LINK_WIDTH_2X_SUP 277 IB_PORT_LINK_WIDTH_2X_SUP = 1 << 4, 278 IB_PORT_LINK_SPEED_HDR_SUP 278 IB_PORT_LINK_SPEED_HDR_SUP = 1 << 5, 279 IB_PORT_LINK_SPEED_NDR_SUP 279 IB_PORT_LINK_SPEED_NDR_SUP = 1 << 10, 280 IB_PORT_EXTENDED_SPEEDS2_SUP << 281 IB_PORT_LINK_SPEED_XDR_SUP << 282 }; 280 }; 283 281 284 #define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26) 282 #define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26) 285 283 286 struct opa_class_port_info { 284 struct opa_class_port_info { 287 u8 base_version; 285 u8 base_version; 288 u8 class_version; 286 u8 class_version; 289 __be16 cap_mask; 287 __be16 cap_mask; 290 __be32 cap_mask2_resp_time; 288 __be32 cap_mask2_resp_time; 291 289 292 u8 redirect_gid[16]; 290 u8 redirect_gid[16]; 293 __be32 redirect_tc_fl; 291 __be32 redirect_tc_fl; 294 __be32 redirect_lid; 292 __be32 redirect_lid; 295 __be32 redirect_sl_qp; 293 __be32 redirect_sl_qp; 296 __be32 redirect_qkey; 294 __be32 redirect_qkey; 297 295 298 u8 trap_gid[16]; 296 u8 trap_gid[16]; 299 __be32 trap_tc_fl; 297 __be32 trap_tc_fl; 300 __be32 trap_lid; 298 __be32 trap_lid; 301 __be32 trap_hl_qp; 299 __be32 trap_hl_qp; 302 __be32 trap_qkey; 300 __be32 trap_qkey; 303 301 304 __be16 trap_pkey; 302 __be16 trap_pkey; 305 __be16 redirect_pkey; 303 __be16 redirect_pkey; 306 304 307 u8 trap_sl_rsvd; 305 u8 trap_sl_rsvd; 308 u8 reserved[3]; 306 u8 reserved[3]; 309 } __packed; 307 } __packed; 310 308 311 /** 309 /** 312 * ib_get_cpi_resp_time - Returns the resp_tim 310 * ib_get_cpi_resp_time - Returns the resp_time value from 313 * cap_mask2_resp_time in ib_class_port_info. 311 * cap_mask2_resp_time in ib_class_port_info. 314 * @cpi: A struct ib_class_port_info mad. 312 * @cpi: A struct ib_class_port_info mad. 315 */ 313 */ 316 static inline u8 ib_get_cpi_resp_time(struct i 314 static inline u8 ib_get_cpi_resp_time(struct ib_class_port_info *cpi) 317 { 315 { 318 return (u8)(be32_to_cpu(cpi->cap_mask2 316 return (u8)(be32_to_cpu(cpi->cap_mask2_resp_time) & 319 IB_CLASS_PORT_INFO_RESP_TI 317 IB_CLASS_PORT_INFO_RESP_TIME_MASK); 320 } 318 } 321 319 322 /** 320 /** 323 * ib_set_cpi_resptime - Sets the response tim 321 * ib_set_cpi_resptime - Sets the response time in an 324 * ib_class_port_info mad. 322 * ib_class_port_info mad. 325 * @cpi: A struct ib_class_port_info. 323 * @cpi: A struct ib_class_port_info. 326 * @rtime: The response time to set. 324 * @rtime: The response time to set. 327 */ 325 */ 328 static inline void ib_set_cpi_resp_time(struct 326 static inline void ib_set_cpi_resp_time(struct ib_class_port_info *cpi, 329 u8 rti 327 u8 rtime) 330 { 328 { 331 cpi->cap_mask2_resp_time = 329 cpi->cap_mask2_resp_time = 332 (cpi->cap_mask2_resp_time & 330 (cpi->cap_mask2_resp_time & 333 cpu_to_be32(~IB_CLASS_PORT_IN 331 cpu_to_be32(~IB_CLASS_PORT_INFO_RESP_TIME_MASK)) | 334 cpu_to_be32(rtime & IB_CLASS_P 332 cpu_to_be32(rtime & IB_CLASS_PORT_INFO_RESP_TIME_MASK); 335 } 333 } 336 334 337 /** 335 /** 338 * ib_get_cpi_capmask2 - Returns the capmask2 336 * ib_get_cpi_capmask2 - Returns the capmask2 value from 339 * cap_mask2_resp_time in ib_class_port_info. 337 * cap_mask2_resp_time in ib_class_port_info. 340 * @cpi: A struct ib_class_port_info mad. 338 * @cpi: A struct ib_class_port_info mad. 341 */ 339 */ 342 static inline u32 ib_get_cpi_capmask2(struct i 340 static inline u32 ib_get_cpi_capmask2(struct ib_class_port_info *cpi) 343 { 341 { 344 return (be32_to_cpu(cpi->cap_mask2_res 342 return (be32_to_cpu(cpi->cap_mask2_resp_time) >> 345 IB_CLASS_PORT_INFO_RESP_TIME_F 343 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 346 } 344 } 347 345 348 /** 346 /** 349 * ib_set_cpi_capmask2 - Sets the capmask2 in 347 * ib_set_cpi_capmask2 - Sets the capmask2 in an 350 * ib_class_port_info mad. 348 * ib_class_port_info mad. 351 * @cpi: A struct ib_class_port_info. 349 * @cpi: A struct ib_class_port_info. 352 * @capmask2: The capmask2 to set. 350 * @capmask2: The capmask2 to set. 353 */ 351 */ 354 static inline void ib_set_cpi_capmask2(struct 352 static inline void ib_set_cpi_capmask2(struct ib_class_port_info *cpi, 355 u32 cap 353 u32 capmask2) 356 { 354 { 357 cpi->cap_mask2_resp_time = 355 cpi->cap_mask2_resp_time = 358 (cpi->cap_mask2_resp_time & 356 (cpi->cap_mask2_resp_time & 359 cpu_to_be32(IB_CLASS_PORT_INF 357 cpu_to_be32(IB_CLASS_PORT_INFO_RESP_TIME_MASK)) | 360 cpu_to_be32(capmask2 << 358 cpu_to_be32(capmask2 << 361 IB_CLASS_PORT_INFO 359 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 362 } 360 } 363 361 364 /** 362 /** 365 * opa_get_cpi_capmask2 - Returns the capmask2 363 * opa_get_cpi_capmask2 - Returns the capmask2 value from 366 * cap_mask2_resp_time in ib_class_port_info. 364 * cap_mask2_resp_time in ib_class_port_info. 367 * @cpi: A struct opa_class_port_info mad. 365 * @cpi: A struct opa_class_port_info mad. 368 */ 366 */ 369 static inline u32 opa_get_cpi_capmask2(struct 367 static inline u32 opa_get_cpi_capmask2(struct opa_class_port_info *cpi) 370 { 368 { 371 return (be32_to_cpu(cpi->cap_mask2_res 369 return (be32_to_cpu(cpi->cap_mask2_resp_time) >> 372 IB_CLASS_PORT_INFO_RESP_TIME_F 370 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 373 } 371 } 374 372 375 struct ib_mad_notice_attr { 373 struct ib_mad_notice_attr { 376 u8 generic_type; 374 u8 generic_type; 377 u8 prod_type_msb; 375 u8 prod_type_msb; 378 __be16 prod_type_lsb; 376 __be16 prod_type_lsb; 379 __be16 trap_num; 377 __be16 trap_num; 380 __be16 issuer_lid; 378 __be16 issuer_lid; 381 __be16 toggle_count; 379 __be16 toggle_count; 382 380 383 union { 381 union { 384 struct { 382 struct { 385 u8 details[54]; 383 u8 details[54]; 386 } raw_data; 384 } raw_data; 387 385 388 struct { 386 struct { 389 __be16 reserved; 387 __be16 reserved; 390 __be16 lid; 388 __be16 lid; /* where violation happened */ 391 u8 port_num; 389 u8 port_num; /* where violation happened */ 392 } __packed ntc_129_131; 390 } __packed ntc_129_131; 393 391 394 struct { 392 struct { 395 __be16 reserved; 393 __be16 reserved; 396 __be16 lid; 394 __be16 lid; /* LID where change occurred */ 397 u8 reserved2; 395 u8 reserved2; 398 u8 local_changes; 396 u8 local_changes; /* low bit - local changes */ 399 __be32 new_cap_mask; 397 __be32 new_cap_mask; /* new capability mask */ 400 u8 reserved3; 398 u8 reserved3; 401 u8 change_flags; 399 u8 change_flags; /* low 3 bits only */ 402 } __packed ntc_144; 400 } __packed ntc_144; 403 401 404 struct { 402 struct { 405 __be16 reserved; 403 __be16 reserved; 406 __be16 lid; 404 __be16 lid; /* lid where sys guid changed */ 407 __be16 reserved2; 405 __be16 reserved2; 408 __be64 new_sys_guid; 406 __be64 new_sys_guid; 409 } __packed ntc_145; 407 } __packed ntc_145; 410 408 411 struct { 409 struct { 412 __be16 reserved; 410 __be16 reserved; 413 __be16 lid; 411 __be16 lid; 414 __be16 dr_slid; 412 __be16 dr_slid; 415 u8 method; 413 u8 method; 416 u8 reserved2; 414 u8 reserved2; 417 __be16 attr_id; 415 __be16 attr_id; 418 __be32 attr_mod; 416 __be32 attr_mod; 419 __be64 mkey; 417 __be64 mkey; 420 u8 reserved3; 418 u8 reserved3; 421 u8 dr_trunc_hop; 419 u8 dr_trunc_hop; 422 u8 dr_rtn_path[30 420 u8 dr_rtn_path[30]; 423 } __packed ntc_256; 421 } __packed ntc_256; 424 422 425 struct { 423 struct { 426 __be16 reserv 424 __be16 reserved; 427 __be16 lid1; 425 __be16 lid1; 428 __be16 lid2; 426 __be16 lid2; 429 __be32 key; 427 __be32 key; 430 __be32 sl_qp1 428 __be32 sl_qp1; /* SL: high 4 bits */ 431 __be32 qp2; 429 __be32 qp2; /* high 8 bits reserved */ 432 union ib_gid gid1; 430 union ib_gid gid1; 433 union ib_gid gid2; 431 union ib_gid gid2; 434 } __packed ntc_257_258; 432 } __packed ntc_257_258; 435 433 436 } details; 434 } details; 437 }; 435 }; 438 436 439 /** 437 /** 440 * ib_mad_send_buf - MAD data buffer and work 438 * ib_mad_send_buf - MAD data buffer and work request for sends. 441 * @next: A pointer used to chain together MAD 439 * @next: A pointer used to chain together MADs for posting. 442 * @mad: References an allocated MAD data buff 440 * @mad: References an allocated MAD data buffer for MADs that do not have 443 * RMPP active. For MADs using RMPP, refere 441 * RMPP active. For MADs using RMPP, references the common and management 444 * class specific headers. 442 * class specific headers. 445 * @mad_agent: MAD agent that allocated the bu 443 * @mad_agent: MAD agent that allocated the buffer. 446 * @ah: The address handle to use when sending 444 * @ah: The address handle to use when sending the MAD. 447 * @context: User-controlled context fields. 445 * @context: User-controlled context fields. 448 * @hdr_len: Indicates the size of the data he 446 * @hdr_len: Indicates the size of the data header of the MAD. This length 449 * includes the common MAD, RMPP, and class 447 * includes the common MAD, RMPP, and class specific headers. 450 * @data_len: Indicates the total size of user 448 * @data_len: Indicates the total size of user-transferred data. 451 * @seg_count: The number of RMPP segments all 449 * @seg_count: The number of RMPP segments allocated for this send. 452 * @seg_size: Size of the data in each RMPP se 450 * @seg_size: Size of the data in each RMPP segment. This does not include 453 * class specific headers. 451 * class specific headers. 454 * @seg_rmpp_size: Size of each RMPP segment i 452 * @seg_rmpp_size: Size of each RMPP segment including the class specific 455 * headers. 453 * headers. 456 * @timeout_ms: Time to wait for a response. 454 * @timeout_ms: Time to wait for a response. 457 * @retries: Number of times to retry a reques 455 * @retries: Number of times to retry a request for a response. For MADs 458 * using RMPP, this applies per window. On 456 * using RMPP, this applies per window. On completion, returns the number 459 * of retries needed to complete the transfe 457 * of retries needed to complete the transfer. 460 * 458 * 461 * Users are responsible for initializing the 459 * Users are responsible for initializing the MAD buffer itself, with the 462 * exception of any RMPP header. Additional s 460 * exception of any RMPP header. Additional segment buffer space allocated 463 * beyond data_len is padding. 461 * beyond data_len is padding. 464 */ 462 */ 465 struct ib_mad_send_buf { 463 struct ib_mad_send_buf { 466 struct ib_mad_send_buf *next; 464 struct ib_mad_send_buf *next; 467 void *mad; 465 void *mad; 468 struct ib_mad_agent *mad_agent; 466 struct ib_mad_agent *mad_agent; 469 struct ib_ah *ah; 467 struct ib_ah *ah; 470 void *context[2]; 468 void *context[2]; 471 int hdr_len; 469 int hdr_len; 472 int data_len; 470 int data_len; 473 int seg_count; 471 int seg_count; 474 int seg_size; 472 int seg_size; 475 int seg_rmpp_size; 473 int seg_rmpp_size; 476 int timeout_ms; 474 int timeout_ms; 477 int retries; 475 int retries; 478 }; 476 }; 479 477 480 /** 478 /** 481 * ib_response_mad - Returns if the specified 479 * ib_response_mad - Returns if the specified MAD has been generated in 482 * response to a sent request or trap. 480 * response to a sent request or trap. 483 */ 481 */ 484 int ib_response_mad(const struct ib_mad_hdr *h 482 int ib_response_mad(const struct ib_mad_hdr *hdr); 485 483 486 /** 484 /** 487 * ib_get_rmpp_resptime - Returns the RMPP res 485 * ib_get_rmpp_resptime - Returns the RMPP response time. 488 * @rmpp_hdr: An RMPP header. 486 * @rmpp_hdr: An RMPP header. 489 */ 487 */ 490 static inline u8 ib_get_rmpp_resptime(struct i 488 static inline u8 ib_get_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr) 491 { 489 { 492 return rmpp_hdr->rmpp_rtime_flags >> 3 490 return rmpp_hdr->rmpp_rtime_flags >> 3; 493 } 491 } 494 492 495 /** 493 /** 496 * ib_get_rmpp_flags - Returns the RMPP flags. 494 * ib_get_rmpp_flags - Returns the RMPP flags. 497 * @rmpp_hdr: An RMPP header. 495 * @rmpp_hdr: An RMPP header. 498 */ 496 */ 499 static inline u8 ib_get_rmpp_flags(struct ib_r 497 static inline u8 ib_get_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr) 500 { 498 { 501 return rmpp_hdr->rmpp_rtime_flags & 0x 499 return rmpp_hdr->rmpp_rtime_flags & 0x7; 502 } 500 } 503 501 504 /** 502 /** 505 * ib_set_rmpp_resptime - Sets the response ti 503 * ib_set_rmpp_resptime - Sets the response time in an RMPP header. 506 * @rmpp_hdr: An RMPP header. 504 * @rmpp_hdr: An RMPP header. 507 * @rtime: The response time to set. 505 * @rtime: The response time to set. 508 */ 506 */ 509 static inline void ib_set_rmpp_resptime(struct 507 static inline void ib_set_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr, u8 rtime) 510 { 508 { 511 rmpp_hdr->rmpp_rtime_flags = ib_get_rm 509 rmpp_hdr->rmpp_rtime_flags = ib_get_rmpp_flags(rmpp_hdr) | (rtime << 3); 512 } 510 } 513 511 514 /** 512 /** 515 * ib_set_rmpp_flags - Sets the flags in an RM 513 * ib_set_rmpp_flags - Sets the flags in an RMPP header. 516 * @rmpp_hdr: An RMPP header. 514 * @rmpp_hdr: An RMPP header. 517 * @flags: The flags to set. 515 * @flags: The flags to set. 518 */ 516 */ 519 static inline void ib_set_rmpp_flags(struct ib 517 static inline void ib_set_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr, u8 flags) 520 { 518 { 521 rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr 519 rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr->rmpp_rtime_flags & 0xF8) | 522 (flags & 520 (flags & 0x7); 523 } 521 } 524 522 525 struct ib_mad_agent; 523 struct ib_mad_agent; 526 struct ib_mad_send_wc; 524 struct ib_mad_send_wc; 527 struct ib_mad_recv_wc; 525 struct ib_mad_recv_wc; 528 526 529 /** 527 /** 530 * ib_mad_send_handler - callback handler for 528 * ib_mad_send_handler - callback handler for a sent MAD. 531 * @mad_agent: MAD agent that sent the MAD. 529 * @mad_agent: MAD agent that sent the MAD. 532 * @mad_send_wc: Send work completion informat 530 * @mad_send_wc: Send work completion information on the sent MAD. 533 */ 531 */ 534 typedef void (*ib_mad_send_handler)(struct ib_ 532 typedef void (*ib_mad_send_handler)(struct ib_mad_agent *mad_agent, 535 struct ib_ 533 struct ib_mad_send_wc *mad_send_wc); 536 534 537 /** 535 /** 538 * ib_mad_recv_handler - callback handler for 536 * ib_mad_recv_handler - callback handler for a received MAD. 539 * @mad_agent: MAD agent requesting the receiv 537 * @mad_agent: MAD agent requesting the received MAD. 540 * @send_buf: Send buffer if found, else NULL 538 * @send_buf: Send buffer if found, else NULL 541 * @mad_recv_wc: Received work completion info 539 * @mad_recv_wc: Received work completion information on the received MAD. 542 * 540 * 543 * MADs received in response to a send request 541 * MADs received in response to a send request operation will be handed to 544 * the user before the send operation complete 542 * the user before the send operation completes. All data buffers given 545 * to registered agents through this routine a 543 * to registered agents through this routine are owned by the receiving 546 * client. 544 * client. 547 */ 545 */ 548 typedef void (*ib_mad_recv_handler)(struct ib_ 546 typedef void (*ib_mad_recv_handler)(struct ib_mad_agent *mad_agent, 549 struct ib_ 547 struct ib_mad_send_buf *send_buf, 550 struct ib_ 548 struct ib_mad_recv_wc *mad_recv_wc); 551 549 552 /** 550 /** 553 * ib_mad_agent - Used to track MAD registrati 551 * ib_mad_agent - Used to track MAD registration with the access layer. 554 * @device: Reference to device registration i 552 * @device: Reference to device registration is on. 555 * @qp: Reference to QP used for sending and r 553 * @qp: Reference to QP used for sending and receiving MADs. 556 * @mr: Memory region for system memory usable 554 * @mr: Memory region for system memory usable for DMA. 557 * @recv_handler: Callback handler for a recei 555 * @recv_handler: Callback handler for a received MAD. 558 * @send_handler: Callback handler for a sent 556 * @send_handler: Callback handler for a sent MAD. 559 * @context: User-specified context associated 557 * @context: User-specified context associated with this registration. 560 * @hi_tid: Access layer assigned transaction 558 * @hi_tid: Access layer assigned transaction ID for this client. 561 * Unsolicited MADs sent by this client will 559 * Unsolicited MADs sent by this client will have the upper 32-bits 562 * of their TID set to this value. 560 * of their TID set to this value. 563 * @flags: registration flags 561 * @flags: registration flags 564 * @port_num: Port number on which QP is regis 562 * @port_num: Port number on which QP is registered 565 * @rmpp_version: If set, indicates the RMPP v 563 * @rmpp_version: If set, indicates the RMPP version used by this agent. 566 */ 564 */ 567 enum { 565 enum { 568 IB_MAD_USER_RMPP = IB_USER_MAD_USER_RM 566 IB_MAD_USER_RMPP = IB_USER_MAD_USER_RMPP, 569 }; 567 }; 570 struct ib_mad_agent { 568 struct ib_mad_agent { 571 struct ib_device *device; 569 struct ib_device *device; 572 struct ib_qp *qp; 570 struct ib_qp *qp; 573 ib_mad_recv_handler recv_handler; 571 ib_mad_recv_handler recv_handler; 574 ib_mad_send_handler send_handler; 572 ib_mad_send_handler send_handler; 575 void *context; 573 void *context; 576 u32 hi_tid; 574 u32 hi_tid; 577 u32 flags; 575 u32 flags; 578 void *security; 576 void *security; 579 struct list_head mad_agent_sec_ 577 struct list_head mad_agent_sec_list; 580 u8 port_num; 578 u8 port_num; 581 u8 rmpp_version; 579 u8 rmpp_version; 582 bool smp_allowed; 580 bool smp_allowed; 583 }; 581 }; 584 582 585 /** 583 /** 586 * ib_mad_send_wc - MAD send completion inform 584 * ib_mad_send_wc - MAD send completion information. 587 * @send_buf: Send MAD data buffer associated 585 * @send_buf: Send MAD data buffer associated with the send MAD request. 588 * @status: Completion status. 586 * @status: Completion status. 589 * @vendor_err: Optional vendor error informat 587 * @vendor_err: Optional vendor error information returned with a failed 590 * request. 588 * request. 591 */ 589 */ 592 struct ib_mad_send_wc { 590 struct ib_mad_send_wc { 593 struct ib_mad_send_buf *send_buf; 591 struct ib_mad_send_buf *send_buf; 594 enum ib_wc_status status; 592 enum ib_wc_status status; 595 u32 vendor_err; 593 u32 vendor_err; 596 }; 594 }; 597 595 598 /** 596 /** 599 * ib_mad_recv_buf - received MAD buffer infor 597 * ib_mad_recv_buf - received MAD buffer information. 600 * @list: Reference to next data buffer for a 598 * @list: Reference to next data buffer for a received RMPP MAD. 601 * @grh: References a data buffer containing t 599 * @grh: References a data buffer containing the global route header. 602 * The data refereced by this buffer is only 600 * The data refereced by this buffer is only valid if the GRH is 603 * valid. 601 * valid. 604 * @mad: References the start of the received 602 * @mad: References the start of the received MAD. 605 */ 603 */ 606 struct ib_mad_recv_buf { 604 struct ib_mad_recv_buf { 607 struct list_head list; 605 struct list_head list; 608 struct ib_grh *grh; 606 struct ib_grh *grh; 609 union { 607 union { 610 struct ib_mad *mad; 608 struct ib_mad *mad; 611 struct opa_mad *opa_mad; 609 struct opa_mad *opa_mad; 612 }; 610 }; 613 }; 611 }; 614 612 615 /** 613 /** 616 * ib_mad_recv_wc - received MAD information. 614 * ib_mad_recv_wc - received MAD information. 617 * @wc: Completion information for the receive 615 * @wc: Completion information for the received data. 618 * @recv_buf: Specifies the location of the re 616 * @recv_buf: Specifies the location of the received data buffer(s). 619 * @rmpp_list: Specifies a list of RMPP reasse 617 * @rmpp_list: Specifies a list of RMPP reassembled received MAD buffers. 620 * @mad_len: The length of the received MAD, w 618 * @mad_len: The length of the received MAD, without duplicated headers. 621 * @mad_seg_size: The size of individual MAD s 619 * @mad_seg_size: The size of individual MAD segments 622 * 620 * 623 * For received response, the wr_id contains a 621 * For received response, the wr_id contains a pointer to the ib_mad_send_buf 624 * for the corresponding send request. 622 * for the corresponding send request. 625 */ 623 */ 626 struct ib_mad_recv_wc { 624 struct ib_mad_recv_wc { 627 struct ib_wc *wc; 625 struct ib_wc *wc; 628 struct ib_mad_recv_buf recv_buf; 626 struct ib_mad_recv_buf recv_buf; 629 struct list_head rmpp_list; 627 struct list_head rmpp_list; 630 int mad_len; 628 int mad_len; 631 size_t mad_seg_size; 629 size_t mad_seg_size; 632 }; 630 }; 633 631 634 /** 632 /** 635 * ib_mad_reg_req - MAD registration request 633 * ib_mad_reg_req - MAD registration request 636 * @mgmt_class: Indicates which management cla 634 * @mgmt_class: Indicates which management class of MADs should be receive 637 * by the caller. This field is only requir 635 * by the caller. This field is only required if the user wishes to 638 * receive unsolicited MADs, otherwise it sh 636 * receive unsolicited MADs, otherwise it should be 0. 639 * @mgmt_class_version: Indicates which versio 637 * @mgmt_class_version: Indicates which version of MADs for the given 640 * management class to receive. 638 * management class to receive. 641 * @oui: Indicates IEEE OUI when mgmt_class is 639 * @oui: Indicates IEEE OUI when mgmt_class is a vendor class 642 * in the range from 0x30 to 0x4f. Otherwise 640 * in the range from 0x30 to 0x4f. Otherwise not used. 643 * @method_mask: The caller will receive unsol 641 * @method_mask: The caller will receive unsolicited MADs for any method 644 * where @method_mask = 1. 642 * where @method_mask = 1. 645 * 643 * 646 */ 644 */ 647 struct ib_mad_reg_req { 645 struct ib_mad_reg_req { 648 u8 mgmt_class; 646 u8 mgmt_class; 649 u8 mgmt_class_version; 647 u8 mgmt_class_version; 650 u8 oui[3]; 648 u8 oui[3]; 651 DECLARE_BITMAP(method_mask, IB_MGMT_MA 649 DECLARE_BITMAP(method_mask, IB_MGMT_MAX_METHODS); 652 }; 650 }; 653 651 654 /** 652 /** 655 * ib_register_mad_agent - Register to send/re 653 * ib_register_mad_agent - Register to send/receive MADs. 656 * @device: The device to register with. 654 * @device: The device to register with. 657 * @port_num: The port on the specified device 655 * @port_num: The port on the specified device to use. 658 * @qp_type: Specifies which QP to access. Mu 656 * @qp_type: Specifies which QP to access. Must be either 659 * IB_QPT_SMI or IB_QPT_GSI. 657 * IB_QPT_SMI or IB_QPT_GSI. 660 * @mad_reg_req: Specifies which unsolicited M 658 * @mad_reg_req: Specifies which unsolicited MADs should be received 661 * by the caller. This parameter may be NUL 659 * by the caller. This parameter may be NULL if the caller only 662 * wishes to receive solicited responses. 660 * wishes to receive solicited responses. 663 * @rmpp_version: If set, indicates that the c 661 * @rmpp_version: If set, indicates that the client will send 664 * and receive MADs that contain the RMPP he 662 * and receive MADs that contain the RMPP header for the given version. 665 * If set to 0, indicates that RMPP is not u 663 * If set to 0, indicates that RMPP is not used by this client. 666 * @send_handler: The completion callback rout 664 * @send_handler: The completion callback routine invoked after a send 667 * request has completed. 665 * request has completed. 668 * @recv_handler: The completion callback rout 666 * @recv_handler: The completion callback routine invoked for a received 669 * MAD. 667 * MAD. 670 * @context: User specified context associated 668 * @context: User specified context associated with the registration. 671 * @registration_flags: Registration flags to 669 * @registration_flags: Registration flags to set for this agent 672 */ 670 */ 673 struct ib_mad_agent *ib_register_mad_agent(str 671 struct ib_mad_agent *ib_register_mad_agent(struct ib_device *device, 674 u32 672 u32 port_num, 675 enu 673 enum ib_qp_type qp_type, 676 str 674 struct ib_mad_reg_req *mad_reg_req, 677 u8 675 u8 rmpp_version, 678 ib_ 676 ib_mad_send_handler send_handler, 679 ib_ 677 ib_mad_recv_handler recv_handler, 680 voi 678 void *context, 681 u32 679 u32 registration_flags); 682 /** 680 /** 683 * ib_unregister_mad_agent - Unregisters a cli 681 * ib_unregister_mad_agent - Unregisters a client from using MAD services. 684 * @mad_agent: Corresponding MAD registration 682 * @mad_agent: Corresponding MAD registration request to deregister. 685 * 683 * 686 * After invoking this routine, MAD services a 684 * After invoking this routine, MAD services are no longer usable by the 687 * client on the associated QP. 685 * client on the associated QP. 688 */ 686 */ 689 void ib_unregister_mad_agent(struct ib_mad_age 687 void ib_unregister_mad_agent(struct ib_mad_agent *mad_agent); 690 688 691 /** 689 /** 692 * ib_post_send_mad - Posts MAD(s) to the send 690 * ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated 693 * with the registered client. 691 * with the registered client. 694 * @send_buf: Specifies the information needed 692 * @send_buf: Specifies the information needed to send the MAD(s). 695 * @bad_send_buf: Specifies the MAD on which a 693 * @bad_send_buf: Specifies the MAD on which an error was encountered. This 696 * parameter is optional if only a single MA 694 * parameter is optional if only a single MAD is posted. 697 * 695 * 698 * Sent MADs are not guaranteed to complete in 696 * Sent MADs are not guaranteed to complete in the order that they were posted. 699 * 697 * 700 * If the MAD requires RMPP, the data buffer s 698 * If the MAD requires RMPP, the data buffer should contain a single copy 701 * of the common MAD, RMPP, and class specific 699 * of the common MAD, RMPP, and class specific headers, followed by the class 702 * defined data. If the class defined data wo 700 * defined data. If the class defined data would not divide evenly into 703 * RMPP segments, then space must be allocated 701 * RMPP segments, then space must be allocated at the end of the referenced 704 * buffer for any required padding. To indica 702 * buffer for any required padding. To indicate the amount of class defined 705 * data being transferred, the paylen_newwin f 703 * data being transferred, the paylen_newwin field in the RMPP header should 706 * be set to the size of the class specific he 704 * be set to the size of the class specific header plus the amount of class 707 * defined data being transferred. The paylen 705 * defined data being transferred. The paylen_newwin field should be 708 * specified in network-byte order. 706 * specified in network-byte order. 709 */ 707 */ 710 int ib_post_send_mad(struct ib_mad_send_buf *s 708 int ib_post_send_mad(struct ib_mad_send_buf *send_buf, 711 struct ib_mad_send_buf ** 709 struct ib_mad_send_buf **bad_send_buf); 712 710 713 711 714 /** 712 /** 715 * ib_free_recv_mad - Returns data buffers use 713 * ib_free_recv_mad - Returns data buffers used to receive a MAD. 716 * @mad_recv_wc: Work completion information f 714 * @mad_recv_wc: Work completion information for a received MAD. 717 * 715 * 718 * Clients receiving MADs through their ib_mad 716 * Clients receiving MADs through their ib_mad_recv_handler must call this 719 * routine to return the work completion buffe 717 * routine to return the work completion buffers to the access layer. 720 */ 718 */ 721 void ib_free_recv_mad(struct ib_mad_recv_wc *m 719 void ib_free_recv_mad(struct ib_mad_recv_wc *mad_recv_wc); 722 720 723 /** 721 /** 724 * ib_modify_mad - Modifies an outstanding sen 722 * ib_modify_mad - Modifies an outstanding send MAD operation. 725 * @send_buf: Indicates the MAD to modify. 723 * @send_buf: Indicates the MAD to modify. 726 * @timeout_ms: New timeout value for sent MAD 724 * @timeout_ms: New timeout value for sent MAD. 727 * 725 * 728 * This call will reset the timeout value for 726 * This call will reset the timeout value for a sent MAD to the specified 729 * value. 727 * value. 730 */ 728 */ 731 int ib_modify_mad(struct ib_mad_send_buf *send 729 int ib_modify_mad(struct ib_mad_send_buf *send_buf, u32 timeout_ms); 732 730 733 /** 731 /** 734 * ib_cancel_mad - Cancels an outstanding send 732 * ib_cancel_mad - Cancels an outstanding send MAD operation. 735 * @send_buf: Indicates the MAD to cancel. 733 * @send_buf: Indicates the MAD to cancel. 736 * 734 * 737 * MADs will be returned to the user through t 735 * MADs will be returned to the user through the corresponding 738 * ib_mad_send_handler. 736 * ib_mad_send_handler. 739 */ 737 */ 740 static inline void ib_cancel_mad(struct ib_mad 738 static inline void ib_cancel_mad(struct ib_mad_send_buf *send_buf) 741 { 739 { 742 ib_modify_mad(send_buf, 0); 740 ib_modify_mad(send_buf, 0); 743 } 741 } 744 742 745 /** 743 /** 746 * ib_create_send_mad - Allocate and initializ 744 * ib_create_send_mad - Allocate and initialize a data buffer and work request 747 * for sending a MAD. 745 * for sending a MAD. 748 * @mad_agent: Specifies the registered MAD se 746 * @mad_agent: Specifies the registered MAD service to associate with the MAD. 749 * @remote_qpn: Specifies the QPN of the recei 747 * @remote_qpn: Specifies the QPN of the receiving node. 750 * @pkey_index: Specifies which PKey the MAD w 748 * @pkey_index: Specifies which PKey the MAD will be sent using. This field 751 * is valid only if the remote_qpn is QP 1. 749 * is valid only if the remote_qpn is QP 1. 752 * @rmpp_active: Indicates if the send will en 750 * @rmpp_active: Indicates if the send will enable RMPP. 753 * @hdr_len: Indicates the size of the data he 751 * @hdr_len: Indicates the size of the data header of the MAD. This length 754 * should include the common MAD header, RMP 752 * should include the common MAD header, RMPP header, plus any class 755 * specific header. 753 * specific header. 756 * @data_len: Indicates the size of any user-t 754 * @data_len: Indicates the size of any user-transferred data. The call will 757 * automatically adjust the allocated buffer 755 * automatically adjust the allocated buffer size to account for any 758 * additional padding that may be necessary. 756 * additional padding that may be necessary. 759 * @gfp_mask: GFP mask used for the memory all 757 * @gfp_mask: GFP mask used for the memory allocation. 760 * @base_version: Base Version of this MAD 758 * @base_version: Base Version of this MAD 761 * 759 * 762 * This routine allocates a MAD for sending. 760 * This routine allocates a MAD for sending. The returned MAD send buffer 763 * will reference a data buffer usable for sen 761 * will reference a data buffer usable for sending a MAD, along 764 * with an initialized work request structure. 762 * with an initialized work request structure. Users may modify the returned 765 * MAD data buffer before posting the send. 763 * MAD data buffer before posting the send. 766 * 764 * 767 * The returned MAD header, class specific hea 765 * The returned MAD header, class specific headers, and any padding will be 768 * cleared. Users are responsible for initial 766 * cleared. Users are responsible for initializing the common MAD header, 769 * any class specific header, and MAD data are 767 * any class specific header, and MAD data area. 770 * If @rmpp_active is set, the RMPP header wil 768 * If @rmpp_active is set, the RMPP header will be initialized for sending. 771 */ 769 */ 772 struct ib_mad_send_buf *ib_create_send_mad(str 770 struct ib_mad_send_buf *ib_create_send_mad(struct ib_mad_agent *mad_agent, 773 u32 771 u32 remote_qpn, u16 pkey_index, 774 int 772 int rmpp_active, 775 int 773 int hdr_len, int data_len, 776 gfp 774 gfp_t gfp_mask, 777 u8 775 u8 base_version); 778 776 779 /** 777 /** 780 * ib_is_mad_class_rmpp - returns whether give 778 * ib_is_mad_class_rmpp - returns whether given management class 781 * supports RMPP. 779 * supports RMPP. 782 * @mgmt_class: management class 780 * @mgmt_class: management class 783 * 781 * 784 * This routine returns whether the management 782 * This routine returns whether the management class supports RMPP. 785 */ 783 */ 786 int ib_is_mad_class_rmpp(u8 mgmt_class); 784 int ib_is_mad_class_rmpp(u8 mgmt_class); 787 785 788 /** 786 /** 789 * ib_get_mad_data_offset - returns the data o 787 * ib_get_mad_data_offset - returns the data offset for a given 790 * management class. 788 * management class. 791 * @mgmt_class: management class 789 * @mgmt_class: management class 792 * 790 * 793 * This routine returns the data offset in the 791 * This routine returns the data offset in the MAD for the management 794 * class requested. 792 * class requested. 795 */ 793 */ 796 int ib_get_mad_data_offset(u8 mgmt_class); 794 int ib_get_mad_data_offset(u8 mgmt_class); 797 795 798 /** 796 /** 799 * ib_get_rmpp_segment - returns the data buff 797 * ib_get_rmpp_segment - returns the data buffer for a given RMPP segment. 800 * @send_buf: Previously allocated send data b 798 * @send_buf: Previously allocated send data buffer. 801 * @seg_num: number of segment to return 799 * @seg_num: number of segment to return 802 * 800 * 803 * This routine returns a pointer to the data 801 * This routine returns a pointer to the data buffer of an RMPP MAD. 804 * Users must provide synchronization to @send 802 * Users must provide synchronization to @send_buf around this call. 805 */ 803 */ 806 void *ib_get_rmpp_segment(struct ib_mad_send_b 804 void *ib_get_rmpp_segment(struct ib_mad_send_buf *send_buf, int seg_num); 807 805 808 /** 806 /** 809 * ib_free_send_mad - Returns data buffers use 807 * ib_free_send_mad - Returns data buffers used to send a MAD. 810 * @send_buf: Previously allocated send data b 808 * @send_buf: Previously allocated send data buffer. 811 */ 809 */ 812 void ib_free_send_mad(struct ib_mad_send_buf * 810 void ib_free_send_mad(struct ib_mad_send_buf *send_buf); 813 811 814 /** 812 /** 815 * ib_mad_kernel_rmpp_agent - Returns if the a 813 * ib_mad_kernel_rmpp_agent - Returns if the agent is performing RMPP. 816 * @agent: the agent in question 814 * @agent: the agent in question 817 * @return: true if agent is performing rmpp, 815 * @return: true if agent is performing rmpp, false otherwise. 818 */ 816 */ 819 int ib_mad_kernel_rmpp_agent(const struct ib_m 817 int ib_mad_kernel_rmpp_agent(const struct ib_mad_agent *agent); 820 818 821 #endif /* IB_MAD_H */ 819 #endif /* IB_MAD_H */ 822 820
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.