1219820Sjeff/* 2219820Sjeff * Copyright (c) 2004 Topspin Communications. All rights reserved. 3219820Sjeff * Copyright (c) 2005, 2006, 2007 Cisco Systems, Inc. All rights reserved. 4272407Shselasky * Copyright (c) 2005, 2006, 2007, 2008, 2014 Mellanox Technologies. All rights reserved. 5219820Sjeff * Copyright (c) 2004 Voltaire, Inc. All rights reserved. 6219820Sjeff * 7219820Sjeff * This software is available to you under a choice of one of two 8219820Sjeff * licenses. You may choose to be licensed under the terms of the GNU 9219820Sjeff * General Public License (GPL) Version 2, available from the file 10219820Sjeff * COPYING in the main directory of this source tree, or the 11219820Sjeff * OpenIB.org BSD license below: 12219820Sjeff * 13219820Sjeff * Redistribution and use in source and binary forms, with or 14219820Sjeff * without modification, are permitted provided that the following 15219820Sjeff * conditions are met: 16219820Sjeff * 17219820Sjeff * - Redistributions of source code must retain the above 18219820Sjeff * copyright notice, this list of conditions and the following 19219820Sjeff * disclaimer. 20219820Sjeff * 21219820Sjeff * - Redistributions in binary form must reproduce the above 22219820Sjeff * copyright notice, this list of conditions and the following 23219820Sjeff * disclaimer in the documentation and/or other materials 24219820Sjeff * provided with the distribution. 25219820Sjeff * 26219820Sjeff * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 27219820Sjeff * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 28219820Sjeff * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 29219820Sjeff * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 30219820Sjeff * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 31219820Sjeff * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 32219820Sjeff * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 33219820Sjeff * SOFTWARE. 34219820Sjeff */ 35219820Sjeff 36272407Shselasky#include <linux/types.h> 37272407Shselasky#include <linux/gfp.h> 38272407Shselasky#include <linux/module.h> 39272407Shselasky 40219820Sjeff#include <linux/mlx4/cmd.h> 41219820Sjeff#include <linux/mlx4/qp.h> 42219820Sjeff 43219820Sjeff#include "mlx4.h" 44219820Sjeff#include "icm.h" 45219820Sjeff 46255932Salfred/* 47255932Salfred * QP to support BF should have bits 6,7 cleared 48255932Salfred */ 49255932Salfred#define MLX4_BF_QP_SKIP_MASK 0xc0 50255932Salfred#define MLX4_MAX_BF_QP_RANGE 0x40 51255932Salfred 52219820Sjeffvoid mlx4_qp_event(struct mlx4_dev *dev, u32 qpn, int event_type) 53219820Sjeff{ 54219820Sjeff struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; 55219820Sjeff struct mlx4_qp *qp; 56219820Sjeff 57219820Sjeff spin_lock(&qp_table->lock); 58219820Sjeff 59219820Sjeff qp = __mlx4_qp_lookup(dev, qpn); 60219820Sjeff if (qp) 61219820Sjeff atomic_inc(&qp->refcount); 62219820Sjeff 63219820Sjeff spin_unlock(&qp_table->lock); 64219820Sjeff 65219820Sjeff if (!qp) { 66255932Salfred mlx4_dbg(dev, "Async event for none existent QP %08x\n", qpn); 67219820Sjeff return; 68219820Sjeff } 69219820Sjeff 70219820Sjeff qp->event(qp, event_type); 71219820Sjeff 72219820Sjeff if (atomic_dec_and_test(&qp->refcount)) 73219820Sjeff complete(&qp->free); 74219820Sjeff} 75219820Sjeff 76255932Salfred/* used for INIT/CLOSE port logic */ 77255932Salfredstatic int is_master_qp0(struct mlx4_dev *dev, struct mlx4_qp *qp, int *real_qp0, int *proxy_qp0) 78219820Sjeff{ 79255932Salfred /* this procedure is called after we already know we are on the master */ 80255932Salfred /* qp0 is either the proxy qp0, or the real qp0 */ 81255932Salfred u32 pf_proxy_offset = dev->phys_caps.base_proxy_sqpn + 8 * mlx4_master_func_num(dev); 82255932Salfred *proxy_qp0 = qp->qpn >= pf_proxy_offset && qp->qpn <= pf_proxy_offset + 1; 83255932Salfred 84255932Salfred *real_qp0 = qp->qpn >= dev->phys_caps.base_sqpn && 85255932Salfred qp->qpn <= dev->phys_caps.base_sqpn + 1; 86255932Salfred 87255932Salfred return *real_qp0 || *proxy_qp0; 88255932Salfred} 89255932Salfred 90255932Salfredstatic int __mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, 91255932Salfred enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, 92255932Salfred struct mlx4_qp_context *context, 93255932Salfred enum mlx4_qp_optpar optpar, 94255932Salfred int sqd_event, struct mlx4_qp *qp, int native) 95255932Salfred{ 96219820Sjeff static const u16 op[MLX4_QP_NUM_STATE][MLX4_QP_NUM_STATE] = { 97219820Sjeff [MLX4_QP_STATE_RST] = { 98219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 99219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 100219820Sjeff [MLX4_QP_STATE_INIT] = MLX4_CMD_RST2INIT_QP, 101219820Sjeff }, 102219820Sjeff [MLX4_QP_STATE_INIT] = { 103219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 104219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 105219820Sjeff [MLX4_QP_STATE_INIT] = MLX4_CMD_INIT2INIT_QP, 106219820Sjeff [MLX4_QP_STATE_RTR] = MLX4_CMD_INIT2RTR_QP, 107219820Sjeff }, 108219820Sjeff [MLX4_QP_STATE_RTR] = { 109219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 110219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 111219820Sjeff [MLX4_QP_STATE_RTS] = MLX4_CMD_RTR2RTS_QP, 112219820Sjeff }, 113219820Sjeff [MLX4_QP_STATE_RTS] = { 114219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 115219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 116219820Sjeff [MLX4_QP_STATE_RTS] = MLX4_CMD_RTS2RTS_QP, 117219820Sjeff [MLX4_QP_STATE_SQD] = MLX4_CMD_RTS2SQD_QP, 118219820Sjeff }, 119219820Sjeff [MLX4_QP_STATE_SQD] = { 120219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 121219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 122219820Sjeff [MLX4_QP_STATE_RTS] = MLX4_CMD_SQD2RTS_QP, 123219820Sjeff [MLX4_QP_STATE_SQD] = MLX4_CMD_SQD2SQD_QP, 124219820Sjeff }, 125219820Sjeff [MLX4_QP_STATE_SQER] = { 126219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 127219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 128219820Sjeff [MLX4_QP_STATE_RTS] = MLX4_CMD_SQERR2RTS_QP, 129219820Sjeff }, 130219820Sjeff [MLX4_QP_STATE_ERR] = { 131219820Sjeff [MLX4_QP_STATE_RST] = MLX4_CMD_2RST_QP, 132219820Sjeff [MLX4_QP_STATE_ERR] = MLX4_CMD_2ERR_QP, 133219820Sjeff } 134219820Sjeff }; 135219820Sjeff 136255932Salfred struct mlx4_priv *priv = mlx4_priv(dev); 137219820Sjeff struct mlx4_cmd_mailbox *mailbox; 138219820Sjeff int ret = 0; 139255932Salfred int real_qp0 = 0; 140255932Salfred int proxy_qp0 = 0; 141255932Salfred u8 port; 142219820Sjeff 143219820Sjeff if (cur_state >= MLX4_QP_NUM_STATE || new_state >= MLX4_QP_NUM_STATE || 144219820Sjeff !op[cur_state][new_state]) 145219820Sjeff return -EINVAL; 146219820Sjeff 147255932Salfred if (op[cur_state][new_state] == MLX4_CMD_2RST_QP) { 148255932Salfred ret = mlx4_cmd(dev, 0, qp->qpn, 2, 149255932Salfred MLX4_CMD_2RST_QP, MLX4_CMD_TIME_CLASS_A, native); 150255932Salfred if (mlx4_is_master(dev) && cur_state != MLX4_QP_STATE_ERR && 151255932Salfred cur_state != MLX4_QP_STATE_RST && 152255932Salfred is_master_qp0(dev, qp, &real_qp0, &proxy_qp0)) { 153255932Salfred port = (qp->qpn & 1) + 1; 154255932Salfred if (proxy_qp0) 155255932Salfred priv->mfunc.master.qp0_state[port].proxy_qp0_active = 0; 156255932Salfred else 157255932Salfred priv->mfunc.master.qp0_state[port].qp0_active = 0; 158255932Salfred } 159255932Salfred return ret; 160255932Salfred } 161219820Sjeff 162219820Sjeff mailbox = mlx4_alloc_cmd_mailbox(dev); 163219820Sjeff if (IS_ERR(mailbox)) 164219820Sjeff return PTR_ERR(mailbox); 165219820Sjeff 166219820Sjeff if (cur_state == MLX4_QP_STATE_RST && new_state == MLX4_QP_STATE_INIT) { 167219820Sjeff u64 mtt_addr = mlx4_mtt_addr(dev, mtt); 168219820Sjeff context->mtt_base_addr_h = mtt_addr >> 32; 169219820Sjeff context->mtt_base_addr_l = cpu_to_be32(mtt_addr & 0xffffffff); 170219820Sjeff context->log_page_size = mtt->page_shift - MLX4_ICM_PAGE_SHIFT; 171219820Sjeff } 172219820Sjeff 173219820Sjeff *(__be32 *) mailbox->buf = cpu_to_be32(optpar); 174219820Sjeff memcpy(mailbox->buf + 8, context, sizeof *context); 175219820Sjeff 176219820Sjeff ((struct mlx4_qp_context *) (mailbox->buf + 8))->local_qpn = 177219820Sjeff cpu_to_be32(qp->qpn); 178219820Sjeff 179255932Salfred ret = mlx4_cmd(dev, mailbox->dma, 180255932Salfred qp->qpn | (!!sqd_event << 31), 181219820Sjeff new_state == MLX4_QP_STATE_RST ? 2 : 0, 182255932Salfred op[cur_state][new_state], MLX4_CMD_TIME_CLASS_C, native); 183219820Sjeff 184255932Salfred if (mlx4_is_master(dev) && is_master_qp0(dev, qp, &real_qp0, &proxy_qp0)) { 185255932Salfred port = (qp->qpn & 1) + 1; 186255932Salfred if (cur_state != MLX4_QP_STATE_ERR && 187255932Salfred cur_state != MLX4_QP_STATE_RST && 188255932Salfred new_state == MLX4_QP_STATE_ERR) { 189255932Salfred if (proxy_qp0) 190255932Salfred priv->mfunc.master.qp0_state[port].proxy_qp0_active = 0; 191255932Salfred else 192255932Salfred priv->mfunc.master.qp0_state[port].qp0_active = 0; 193255932Salfred } else if (new_state == MLX4_QP_STATE_RTR) { 194255932Salfred if (proxy_qp0) 195255932Salfred priv->mfunc.master.qp0_state[port].proxy_qp0_active = 1; 196255932Salfred else 197255932Salfred priv->mfunc.master.qp0_state[port].qp0_active = 1; 198255932Salfred } 199255932Salfred } 200255932Salfred 201219820Sjeff mlx4_free_cmd_mailbox(dev, mailbox); 202219820Sjeff return ret; 203219820Sjeff} 204255932Salfred 205255932Salfredint mlx4_qp_modify(struct mlx4_dev *dev, struct mlx4_mtt *mtt, 206255932Salfred enum mlx4_qp_state cur_state, enum mlx4_qp_state new_state, 207255932Salfred struct mlx4_qp_context *context, 208255932Salfred enum mlx4_qp_optpar optpar, 209255932Salfred int sqd_event, struct mlx4_qp *qp) 210255932Salfred{ 211255932Salfred return __mlx4_qp_modify(dev, mtt, cur_state, new_state, context, 212255932Salfred optpar, sqd_event, qp, 0); 213255932Salfred} 214219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_modify); 215219820Sjeff 216255932Salfredint __mlx4_qp_reserve_range(struct mlx4_dev *dev, int cnt, int align, 217272407Shselasky int *base, u8 flags) 218219820Sjeff{ 219272407Shselasky int bf_qp = !!(flags & (u8) MLX4_RESERVE_BF_QP); 220272407Shselasky 221219820Sjeff struct mlx4_priv *priv = mlx4_priv(dev); 222219820Sjeff struct mlx4_qp_table *qp_table = &priv->qp_table; 223219820Sjeff 224272407Shselasky /* Only IPoIB uses a large cnt. In this case, just allocate 225272407Shselasky * as usual, ignoring bf skipping, since IPoIB does not run over RoCE 226272407Shselasky */ 227255932Salfred if (cnt > MLX4_MAX_BF_QP_RANGE && bf_qp) 228272407Shselasky bf_qp = 0; 229219820Sjeff 230255932Salfred *base = mlx4_bitmap_alloc_range(&qp_table->bitmap, cnt, align, 231255932Salfred bf_qp ? MLX4_BF_QP_SKIP_MASK : 0); 232255932Salfred if (*base == -1) 233255932Salfred return -ENOMEM; 234255932Salfred 235219820Sjeff return 0; 236219820Sjeff} 237255932Salfred 238255932Salfredint mlx4_qp_reserve_range(struct mlx4_dev *dev, int cnt, int align, 239272407Shselasky int *base, u8 flags) 240255932Salfred{ 241255932Salfred u64 in_param = 0; 242255932Salfred u64 out_param; 243255932Salfred int err; 244255932Salfred 245318533Shselasky /* Turn off all unsupported QP allocation flags */ 246318533Shselasky flags &= dev->caps.alloc_res_qp_mask; 247318533Shselasky 248255932Salfred if (mlx4_is_mfunc(dev)) { 249272407Shselasky set_param_l(&in_param, (((u32) flags) << 24) | (u32) cnt); 250255932Salfred set_param_h(&in_param, align); 251255932Salfred err = mlx4_cmd_imm(dev, in_param, &out_param, 252255932Salfred RES_QP, RES_OP_RESERVE, 253255932Salfred MLX4_CMD_ALLOC_RES, 254255932Salfred MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED); 255255932Salfred if (err) 256255932Salfred return err; 257255932Salfred 258255932Salfred *base = get_param_l(&out_param); 259255932Salfred return 0; 260255932Salfred } 261272407Shselasky return __mlx4_qp_reserve_range(dev, cnt, align, base, flags); 262255932Salfred} 263219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_reserve_range); 264219820Sjeff 265255932Salfredvoid __mlx4_qp_release_range(struct mlx4_dev *dev, int base_qpn, int cnt) 266219820Sjeff{ 267219820Sjeff struct mlx4_priv *priv = mlx4_priv(dev); 268219820Sjeff struct mlx4_qp_table *qp_table = &priv->qp_table; 269255932Salfred 270255932Salfred if (mlx4_is_qp_reserved(dev, (u32) base_qpn)) 271219820Sjeff return; 272272407Shselasky mlx4_bitmap_free_range(&qp_table->bitmap, base_qpn, cnt, MLX4_USE_RR); 273219820Sjeff} 274255932Salfred 275255932Salfredvoid mlx4_qp_release_range(struct mlx4_dev *dev, int base_qpn, int cnt) 276255932Salfred{ 277255932Salfred u64 in_param = 0; 278255932Salfred int err; 279255932Salfred 280255932Salfred if (mlx4_is_mfunc(dev)) { 281255932Salfred set_param_l(&in_param, base_qpn); 282255932Salfred set_param_h(&in_param, cnt); 283255932Salfred err = mlx4_cmd(dev, in_param, RES_QP, RES_OP_RESERVE, 284255932Salfred MLX4_CMD_FREE_RES, 285255932Salfred MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED); 286255932Salfred if (err) { 287255932Salfred mlx4_warn(dev, "Failed to release qp range" 288255932Salfred " base:%d cnt:%d\n", base_qpn, cnt); 289255932Salfred } 290255932Salfred } else 291255932Salfred __mlx4_qp_release_range(dev, base_qpn, cnt); 292255932Salfred} 293219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_release_range); 294219820Sjeff 295255932Salfredint __mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn) 296219820Sjeff{ 297219820Sjeff struct mlx4_priv *priv = mlx4_priv(dev); 298219820Sjeff struct mlx4_qp_table *qp_table = &priv->qp_table; 299219820Sjeff int err; 300219820Sjeff 301255932Salfred err = mlx4_table_get(dev, &qp_table->qp_table, qpn); 302219820Sjeff if (err) 303219820Sjeff goto err_out; 304219820Sjeff 305255932Salfred err = mlx4_table_get(dev, &qp_table->auxc_table, qpn); 306219820Sjeff if (err) 307219820Sjeff goto err_put_qp; 308219820Sjeff 309255932Salfred err = mlx4_table_get(dev, &qp_table->altc_table, qpn); 310219820Sjeff if (err) 311219820Sjeff goto err_put_auxc; 312219820Sjeff 313255932Salfred err = mlx4_table_get(dev, &qp_table->rdmarc_table, qpn); 314219820Sjeff if (err) 315219820Sjeff goto err_put_altc; 316219820Sjeff 317255932Salfred err = mlx4_table_get(dev, &qp_table->cmpt_table, qpn); 318219820Sjeff if (err) 319219820Sjeff goto err_put_rdmarc; 320219820Sjeff 321219820Sjeff return 0; 322219820Sjeff 323219820Sjefferr_put_rdmarc: 324255932Salfred mlx4_table_put(dev, &qp_table->rdmarc_table, qpn); 325219820Sjeff 326219820Sjefferr_put_altc: 327255932Salfred mlx4_table_put(dev, &qp_table->altc_table, qpn); 328219820Sjeff 329219820Sjefferr_put_auxc: 330255932Salfred mlx4_table_put(dev, &qp_table->auxc_table, qpn); 331219820Sjeff 332219820Sjefferr_put_qp: 333255932Salfred mlx4_table_put(dev, &qp_table->qp_table, qpn); 334219820Sjeff 335219820Sjefferr_out: 336219820Sjeff return err; 337219820Sjeff} 338219820Sjeff 339255932Salfredstatic int mlx4_qp_alloc_icm(struct mlx4_dev *dev, int qpn) 340219820Sjeff{ 341255932Salfred u64 param = 0; 342219820Sjeff 343255932Salfred if (mlx4_is_mfunc(dev)) { 344255932Salfred set_param_l(¶m, qpn); 345255932Salfred return mlx4_cmd_imm(dev, param, ¶m, RES_QP, RES_OP_MAP_ICM, 346255932Salfred MLX4_CMD_ALLOC_RES, MLX4_CMD_TIME_CLASS_A, 347255932Salfred MLX4_CMD_WRAPPED); 348255932Salfred } 349255932Salfred return __mlx4_qp_alloc_icm(dev, qpn); 350219820Sjeff} 351219820Sjeff 352255932Salfredvoid __mlx4_qp_free_icm(struct mlx4_dev *dev, int qpn) 353255932Salfred{ 354255932Salfred struct mlx4_priv *priv = mlx4_priv(dev); 355255932Salfred struct mlx4_qp_table *qp_table = &priv->qp_table; 356255932Salfred 357255932Salfred mlx4_table_put(dev, &qp_table->cmpt_table, qpn); 358255932Salfred mlx4_table_put(dev, &qp_table->rdmarc_table, qpn); 359255932Salfred mlx4_table_put(dev, &qp_table->altc_table, qpn); 360255932Salfred mlx4_table_put(dev, &qp_table->auxc_table, qpn); 361255932Salfred mlx4_table_put(dev, &qp_table->qp_table, qpn); 362255932Salfred} 363255932Salfred 364255932Salfredstatic void mlx4_qp_free_icm(struct mlx4_dev *dev, int qpn) 365255932Salfred{ 366255932Salfred u64 in_param = 0; 367255932Salfred 368255932Salfred if (mlx4_is_mfunc(dev)) { 369255932Salfred set_param_l(&in_param, qpn); 370255932Salfred if (mlx4_cmd(dev, in_param, RES_QP, RES_OP_MAP_ICM, 371255932Salfred MLX4_CMD_FREE_RES, MLX4_CMD_TIME_CLASS_A, 372255932Salfred MLX4_CMD_WRAPPED)) 373255932Salfred mlx4_warn(dev, "Failed to free icm of qp:%d\n", qpn); 374255932Salfred } else 375255932Salfred __mlx4_qp_free_icm(dev, qpn); 376255932Salfred} 377255932Salfred 378255932Salfredint mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp) 379255932Salfred{ 380255932Salfred struct mlx4_priv *priv = mlx4_priv(dev); 381255932Salfred struct mlx4_qp_table *qp_table = &priv->qp_table; 382255932Salfred int err; 383255932Salfred 384255932Salfred if (!qpn) 385255932Salfred return -EINVAL; 386255932Salfred 387255932Salfred qp->qpn = qpn; 388255932Salfred 389255932Salfred err = mlx4_qp_alloc_icm(dev, qpn); 390255932Salfred if (err) 391255932Salfred return err; 392255932Salfred 393255932Salfred spin_lock_irq(&qp_table->lock); 394255932Salfred err = radix_tree_insert(&dev->qp_table_tree, qp->qpn & 395255932Salfred (dev->caps.num_qps - 1), qp); 396255932Salfred spin_unlock_irq(&qp_table->lock); 397255932Salfred if (err) 398255932Salfred goto err_icm; 399255932Salfred 400255932Salfred atomic_set(&qp->refcount, 1); 401255932Salfred init_completion(&qp->free); 402255932Salfred 403255932Salfred return 0; 404255932Salfred 405255932Salfrederr_icm: 406255932Salfred mlx4_qp_free_icm(dev, qpn); 407255932Salfred return err; 408255932Salfred} 409255932Salfred 410255932SalfredEXPORT_SYMBOL_GPL(mlx4_qp_alloc); 411255932Salfred 412219820Sjeffvoid mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp) 413219820Sjeff{ 414219820Sjeff struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; 415219820Sjeff unsigned long flags; 416219820Sjeff 417219820Sjeff spin_lock_irqsave(&qp_table->lock, flags); 418219820Sjeff radix_tree_delete(&dev->qp_table_tree, qp->qpn & (dev->caps.num_qps - 1)); 419219820Sjeff spin_unlock_irqrestore(&qp_table->lock, flags); 420219820Sjeff} 421219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_remove); 422219820Sjeff 423219820Sjeffvoid mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) 424219820Sjeff{ 425219820Sjeff if (atomic_dec_and_test(&qp->refcount)) 426219820Sjeff complete(&qp->free); 427219820Sjeff wait_for_completion(&qp->free); 428219820Sjeff 429255932Salfred mlx4_qp_free_icm(dev, qp->qpn); 430219820Sjeff} 431219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_free); 432219820Sjeff 433219820Sjeffstatic int mlx4_CONF_SPECIAL_QP(struct mlx4_dev *dev, u32 base_qpn) 434219820Sjeff{ 435255932Salfred return mlx4_cmd(dev, 0, base_qpn, 0, MLX4_CMD_CONF_SPECIAL_QP, 436255932Salfred MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE); 437219820Sjeff} 438219820Sjeff 439219820Sjeffint mlx4_init_qp_table(struct mlx4_dev *dev) 440219820Sjeff{ 441219820Sjeff struct mlx4_qp_table *qp_table = &mlx4_priv(dev)->qp_table; 442219820Sjeff int err; 443219820Sjeff int reserved_from_top = 0; 444255932Salfred int reserved_from_bot; 445255932Salfred int k; 446219820Sjeff 447219820Sjeff spin_lock_init(&qp_table->lock); 448219820Sjeff INIT_RADIX_TREE(&dev->qp_table_tree, GFP_ATOMIC); 449255932Salfred if (mlx4_is_slave(dev)) 450255932Salfred return 0; 451219820Sjeff 452219820Sjeff /* 453219820Sjeff * We reserve 2 extra QPs per port for the special QPs. The 454219820Sjeff * block of special QPs must be aligned to a multiple of 8, so 455219820Sjeff * round up. 456255932Salfred * 457219820Sjeff * We also reserve the MSB of the 24-bit QP number to indicate 458255932Salfred * that a QP is an XRC QP. 459219820Sjeff */ 460255932Salfred dev->phys_caps.base_sqpn = 461219820Sjeff ALIGN(dev->caps.reserved_qps_cnt[MLX4_QP_REGION_FW], 8); 462219820Sjeff 463219820Sjeff { 464219820Sjeff int sort[MLX4_NUM_QP_REGION]; 465219820Sjeff int i, j, tmp; 466219820Sjeff int last_base = dev->caps.num_qps; 467219820Sjeff 468219820Sjeff for (i = 1; i < MLX4_NUM_QP_REGION; ++i) 469219820Sjeff sort[i] = i; 470219820Sjeff 471219820Sjeff for (i = MLX4_NUM_QP_REGION; i > 0; --i) { 472219820Sjeff for (j = 2; j < i; ++j) { 473219820Sjeff if (dev->caps.reserved_qps_cnt[sort[j]] > 474219820Sjeff dev->caps.reserved_qps_cnt[sort[j - 1]]) { 475219820Sjeff tmp = sort[j]; 476219820Sjeff sort[j] = sort[j - 1]; 477219820Sjeff sort[j - 1] = tmp; 478219820Sjeff } 479219820Sjeff } 480219820Sjeff } 481219820Sjeff 482219820Sjeff for (i = 1; i < MLX4_NUM_QP_REGION; ++i) { 483219820Sjeff last_base -= dev->caps.reserved_qps_cnt[sort[i]]; 484219820Sjeff dev->caps.reserved_qps_base[sort[i]] = last_base; 485219820Sjeff reserved_from_top += 486219820Sjeff dev->caps.reserved_qps_cnt[sort[i]]; 487219820Sjeff } 488219820Sjeff 489219820Sjeff } 490219820Sjeff 491255932Salfred /* Reserve 8 real SQPs in both native and SRIOV modes. 492255932Salfred * In addition, in SRIOV mode, reserve 8 proxy SQPs per function 493255932Salfred * (for all PFs and VFs), and 8 corresponding tunnel QPs. 494255932Salfred * Each proxy SQP works opposite its own tunnel QP. 495255932Salfred * 496255932Salfred * The QPs are arranged as follows: 497255932Salfred * a. 8 real SQPs 498255932Salfred * b. All the proxy SQPs (8 per function) 499255932Salfred * c. All the tunnel QPs (8 per function) 500255932Salfred */ 501255932Salfred reserved_from_bot = mlx4_num_reserved_sqps(dev); 502255932Salfred if (reserved_from_bot + reserved_from_top > dev->caps.num_qps) { 503255932Salfred mlx4_err(dev, "Number of reserved QPs is higher than number " 504255932Salfred "of QPs, increase the value of log_num_qp\n"); 505255932Salfred return -EINVAL; 506255932Salfred } 507255932Salfred 508219820Sjeff err = mlx4_bitmap_init(&qp_table->bitmap, dev->caps.num_qps, 509255932Salfred (1 << 23) - 1, reserved_from_bot, 510219820Sjeff reserved_from_top); 511219820Sjeff if (err) 512219820Sjeff return err; 513219820Sjeff 514255932Salfred if (mlx4_is_mfunc(dev)) { 515255932Salfred /* for PPF use */ 516255932Salfred dev->phys_caps.base_proxy_sqpn = dev->phys_caps.base_sqpn + 8; 517255932Salfred dev->phys_caps.base_tunnel_sqpn = dev->phys_caps.base_sqpn + 8 + 8 * MLX4_MFUNC_MAX; 518255932Salfred 519255932Salfred /* In mfunc, calculate proxy and tunnel qp offsets for the PF here, 520255932Salfred * since the PF does not call mlx4_slave_caps */ 521255932Salfred dev->caps.qp0_tunnel = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL); 522255932Salfred dev->caps.qp0_proxy = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL); 523255932Salfred dev->caps.qp1_tunnel = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL); 524255932Salfred dev->caps.qp1_proxy = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL); 525255932Salfred 526255932Salfred if (!dev->caps.qp0_tunnel || !dev->caps.qp0_proxy || 527255932Salfred !dev->caps.qp1_tunnel || !dev->caps.qp1_proxy) { 528255932Salfred err = -ENOMEM; 529255932Salfred goto err_mem; 530255932Salfred } 531255932Salfred 532255932Salfred for (k = 0; k < dev->caps.num_ports; k++) { 533255932Salfred dev->caps.qp0_proxy[k] = dev->phys_caps.base_proxy_sqpn + 534255932Salfred 8 * mlx4_master_func_num(dev) + k; 535255932Salfred dev->caps.qp0_tunnel[k] = dev->caps.qp0_proxy[k] + 8 * MLX4_MFUNC_MAX; 536255932Salfred dev->caps.qp1_proxy[k] = dev->phys_caps.base_proxy_sqpn + 537255932Salfred 8 * mlx4_master_func_num(dev) + MLX4_MAX_PORTS + k; 538255932Salfred dev->caps.qp1_tunnel[k] = dev->caps.qp1_proxy[k] + 8 * MLX4_MFUNC_MAX; 539255932Salfred } 540255932Salfred } 541255932Salfred 542255932Salfred 543255932Salfred err = mlx4_CONF_SPECIAL_QP(dev, dev->phys_caps.base_sqpn); 544255932Salfred if (err) 545255932Salfred goto err_mem; 546255932Salfred return 0; 547255932Salfred 548255932Salfrederr_mem: 549255932Salfred kfree(dev->caps.qp0_tunnel); 550255932Salfred kfree(dev->caps.qp0_proxy); 551255932Salfred kfree(dev->caps.qp1_tunnel); 552255932Salfred kfree(dev->caps.qp1_proxy); 553255932Salfred dev->caps.qp0_tunnel = dev->caps.qp0_proxy = 554255932Salfred dev->caps.qp1_tunnel = dev->caps.qp1_proxy = NULL; 555255932Salfred return err; 556219820Sjeff} 557219820Sjeff 558219820Sjeffvoid mlx4_cleanup_qp_table(struct mlx4_dev *dev) 559219820Sjeff{ 560255932Salfred if (mlx4_is_slave(dev)) 561255932Salfred return; 562255932Salfred 563219820Sjeff mlx4_CONF_SPECIAL_QP(dev, 0); 564219820Sjeff mlx4_bitmap_cleanup(&mlx4_priv(dev)->qp_table.bitmap); 565219820Sjeff} 566219820Sjeff 567219820Sjeffint mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp, 568219820Sjeff struct mlx4_qp_context *context) 569219820Sjeff{ 570219820Sjeff struct mlx4_cmd_mailbox *mailbox; 571219820Sjeff int err; 572219820Sjeff 573219820Sjeff mailbox = mlx4_alloc_cmd_mailbox(dev); 574219820Sjeff if (IS_ERR(mailbox)) 575219820Sjeff return PTR_ERR(mailbox); 576219820Sjeff 577219820Sjeff err = mlx4_cmd_box(dev, 0, mailbox->dma, qp->qpn, 0, 578255932Salfred MLX4_CMD_QUERY_QP, MLX4_CMD_TIME_CLASS_A, 579255932Salfred MLX4_CMD_WRAPPED); 580219820Sjeff if (!err) 581219820Sjeff memcpy(context, mailbox->buf + 8, sizeof *context); 582219820Sjeff 583219820Sjeff mlx4_free_cmd_mailbox(dev, mailbox); 584219820Sjeff return err; 585219820Sjeff} 586219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_query); 587219820Sjeff 588219820Sjeffint mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt, 589219820Sjeff struct mlx4_qp_context *context, 590219820Sjeff struct mlx4_qp *qp, enum mlx4_qp_state *qp_state) 591219820Sjeff{ 592219820Sjeff int err; 593219820Sjeff int i; 594219820Sjeff enum mlx4_qp_state states[] = { 595219820Sjeff MLX4_QP_STATE_RST, 596219820Sjeff MLX4_QP_STATE_INIT, 597219820Sjeff MLX4_QP_STATE_RTR, 598219820Sjeff MLX4_QP_STATE_RTS 599219820Sjeff }; 600219820Sjeff 601219820Sjeff for (i = 0; i < ARRAY_SIZE(states) - 1; i++) { 602219820Sjeff context->flags &= cpu_to_be32(~(0xf << 28)); 603219820Sjeff context->flags |= cpu_to_be32(states[i + 1] << 28); 604219820Sjeff err = mlx4_qp_modify(dev, mtt, states[i], states[i + 1], 605219820Sjeff context, 0, 0, qp); 606219820Sjeff if (err) { 607219820Sjeff mlx4_err(dev, "Failed to bring QP to state: " 608219820Sjeff "%d with error: %d\n", 609219820Sjeff states[i + 1], err); 610219820Sjeff return err; 611219820Sjeff } 612219820Sjeff 613219820Sjeff *qp_state = states[i + 1]; 614219820Sjeff } 615219820Sjeff 616219820Sjeff return 0; 617219820Sjeff} 618219820SjeffEXPORT_SYMBOL_GPL(mlx4_qp_to_ready); 619