/drivers/infiniband/hw/ehca/ |
H A D | ehca_eq.c | 54 struct ehca_eq *eq, 64 spin_lock_init(&eq->spinlock); 65 spin_lock_init(&eq->irq_spinlock); 66 eq->is_initialized = 0; 69 ehca_err(ib_dev, "Invalid EQ type %x. eq=%p", type, eq); 73 ehca_err(ib_dev, "EQ length must not be zero. eq=%p", eq); 78 &eq->pf, 81 &eq 53 ehca_create_eq(struct ehca_shca *shca, struct ehca_eq *eq, const enum ehca_eq_type type, const u32 length) argument 155 ehca_poll_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument 167 ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument [all...] |
H A D | ehca_irq.c | 507 tasklet_hi_schedule(&shca->eq.interrupt_task); 551 struct ehca_eq *eq = &shca->eq; local 552 struct ehca_eqe_cache_entry *eqe_cache = eq->eqe_cache; 557 spin_lock(&eq->irq_spinlock); 564 shca->ipz_hca_handle, eq->ist); 577 eqe_cache[eqe_cnt].eqe = ehca_poll_eq(shca, eq); 604 ret = hipz_h_eoi(eq->ist); 614 if (eq->eqe_cache[i].cq) 615 reset_eq_pending(eq [all...] |
H A D | ehca_main.c | 771 ret = ehca_create_eq(shca, &shca->eq, EHCA_EQ, eq_size); 871 ret = ehca_destroy_eq(shca, &shca->eq); 913 ret = ehca_destroy_eq(shca, &shca->eq); 957 if (shca->eq.is_initialized) { 958 /* call deadman proc only if eq ptr does not change */ 959 struct ehca_eq *eq = &shca->eq; local 963 spin_lock_irqsave(&eq->spinlock, flags); 964 q_ofs = eq->ipz_queue.current_q_offset; 965 spin_unlock_irqrestore(&eq [all...] |
H A D | hcp_if.c | 710 struct ehca_eq *eq) 714 ret = hcp_galpas_dtor(&eq->galpas); 716 ehca_gen_err("Could not destruct eq->galpas"); 722 eq->ipz_eq_handle.handle, /* r5 */ 709 hipz_h_destroy_eq(const struct ipz_adapter_handle adapter_handle, struct ehca_eq *eq) argument
|
H A D | ehca_classes.h | 121 struct ehca_eq eq; member in struct:ehca_shca
|
/drivers/clk/spear/ |
H A D | clk-aux-synth.c | 23 * Aux synth gives rate for different values of eq, x and y 49 u8 eq = rtbl[index].eq ? 1 : 2; local 52 (rtbl[index].yscale * eq)) * 10000; 115 val |= (rtbl[i].eq & aux->masks->eq_sel_mask) <<
|
H A D | clk.h | 46 u8 eq; member in struct:aux_rate_tbl
|
/drivers/staging/lustre/lnet/lnet/ |
H A D | lib-eq.c | 36 * lnet/lnet/lib-eq.c 73 lnet_eq_t *eq; local 93 eq = lnet_eq_alloc(); 94 if (eq == NULL) 98 LIBCFS_ALLOC(eq->eq_events, count * sizeof(lnet_event_t)); 99 if (eq->eq_events == NULL) 105 eq->eq_deq_seq = 1; 106 eq->eq_enq_seq = 1; 107 eq->eq_size = count; 108 eq 155 struct lnet_eq *eq; local 210 lnet_eq_enqueue_event(lnet_eq_t *eq, lnet_event_t *ev) argument 239 lnet_eq_dequeue_event(lnet_eq_t *eq, lnet_event_t *ev) argument 410 lnet_eq_t *eq = lnet_handle2eq(&eventqs[i]); local [all...] |
/drivers/input/ |
H A D | input-mt.c | 296 static int adjust_dual(int *begin, int step, int *end, int eq) argument 314 if (c == 0 || (c > 0 && !eq))
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eq.c | 105 static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) argument 107 return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE); 110 static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) argument 112 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); 114 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; 186 static void eq_update_ci(struct mlx5_eq *eq, int arm) argument 188 __be32 __iomem *addr = eq 195 mlx5_eq_int(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument 312 struct mlx5_eq *eq = eq_ptr; local 321 init_eq_buf(struct mlx5_eq *eq) argument 332 mlx5_create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, u8 vecidx, int nent, u64 mask, const char *name, struct mlx5_uar *uar) argument 411 mlx5_destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument 509 mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq, struct mlx5_query_eq_mbox_out *out, int outlen) argument [all...] |
H A D | debugfs.c | 358 static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, argument 372 err = mlx5_core_eq_query(dev, eq, out, sizeof(*out)); 374 mlx5_core_warn(dev, "failed to query eq\n"); 562 int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument 570 &eq->dbg, eq->eqn, eq_fields, 571 ARRAY_SIZE(eq_fields), eq); 573 eq->dbg = NULL; 578 void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument 583 if (eq [all...] |
/drivers/staging/lustre/include/linux/lnet/ |
H A D | lib-lnet.h | 220 lnet_eq_t *eq; local 225 eq = (lnet_eq_t *)lnet_freelist_alloc(&rec->rec_freelist); 228 return eq; 232 lnet_eq_free_locked(lnet_eq_t *eq) argument 238 lnet_freelist_free(&rec->rec_freelist, eq); 242 lnet_eq_free(lnet_eq_t *eq) argument 245 lnet_eq_free_locked(eq); 365 lnet_eq_t *eq; local 367 LIBCFS_ALLOC(eq, sizeof(*eq)); 372 lnet_eq_free(lnet_eq_t *eq) argument 478 lnet_eq2handle(lnet_handle_eq_t *handle, lnet_eq_t *eq) argument [all...] |
/drivers/infiniband/hw/mlx5/ |
H A D | main.c | 68 struct mlx5_eq *eq, *n; local 72 list_for_each_entry_safe(eq, n, &dev->eqs_list, list) { 73 if (eq->index == vector) { 74 *eqn = eq->eqn; 75 *irqn = eq->irqn; 89 struct mlx5_eq *eq, *n; local 99 eq = kzalloc(sizeof(*eq), GFP_KERNEL); 100 if (!eq) { 106 err = mlx5_create_map_eq(dev->mdev, eq, 140 struct mlx5_eq *eq, *n; local [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_eq.c | 173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), 189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); 199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 202 arbel_set_eq_ci(dev, eq, ci); 204 tavor_set_eq_ci(dev, eq, ci); 228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) argument 230 unsigned long off = (entry & (eq 234 next_eqe_sw(struct mthca_eq *eq) argument 260 mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) argument 420 struct mthca_eq *eq = eq_ptr; local 454 struct mthca_eq *eq = eq_ptr; local 465 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument 588 mthca_free_eq(struct mthca_dev *dev, struct mthca_eq *eq) argument [all...] |
H A D | mthca_dev.h | 231 struct mthca_eq eq[MTHCA_NUM_EQ]; member in struct:mthca_eq_table
|
/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 256 struct ehea_eq *eq; local 258 eq = kzalloc(sizeof(*eq), GFP_KERNEL); 259 if (!eq) 262 eq->adapter = adapter; 263 eq->attr.type = type; 264 eq->attr.max_nr_of_eqes = max_nr_of_eqes; 265 eq->attr.eqe_gen = eqe_gen; 266 spin_lock_init(&eq->spinlock); 269 &eq 323 ehea_poll_eq(struct ehea_eq *eq) argument 335 ehea_destroy_eq_res(struct ehea_eq *eq, u64 force) argument 354 ehea_destroy_eq(struct ehea_eq *eq) argument [all...] |
H A D | ehea.h | 366 struct ehea_eq *eq; member in struct:ehea_port_res
|
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | eq.c | 95 static void eq_set_ci(struct mlx4_eq *eq, int req_not) argument 97 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) | 99 eq->doorbell); 104 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, argument 107 /* (entry & (eq->nent - 1)) gives us a cyclic array */ 108 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; 116 return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE; 119 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) argument 121 struct mlx4_eqe *eqe = get_eqe(eq, eq 449 mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq) argument 781 struct mlx4_eq *eq = eq_ptr; local 852 mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq) argument 886 mlx4_create_eq(struct mlx4_dev *dev, int nent, u8 intr, struct mlx4_eq *eq) argument 999 mlx4_free_eq(struct mlx4_dev *dev, struct mlx4_eq *eq) argument [all...] |
/drivers/scsi/be2iscsi/ |
H A D | be_cmds.c | 824 struct be_queue_info *eq, int eq_delay) 829 struct be_dma_mem *q_mem = &eq->dma_mem; 847 __ilog2_u32(eq->len / 256)); 856 eq->id = le16_to_cpu(resp->eq_id); 857 eq->created = true; 945 struct be_queue_info *cq, struct be_queue_info *eq, 974 AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); 989 AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id); 823 beiscsi_cmd_eq_create(struct be_ctrl_info *ctrl, struct be_queue_info *eq, int eq_delay) argument 944 beiscsi_cmd_cq_create(struct be_ctrl_info *ctrl, struct be_queue_info *cq, struct be_queue_info *eq, bool sol_evts, bool no_delay, int coalesce_wm) argument
|
/drivers/block/drbd/ |
H A D | drbd_worker.c | 1113 int err, eq = 0; local 1139 eq = !memcmp(digest, di->digest, digest_size); 1143 if (eq) { 1236 int err, eq = 0; local 1261 eq = !memcmp(digest, di->digest, digest_size); 1272 if (!eq) 1278 eq ? ID_IN_SYNC : ID_OUT_OF_SYNC);
|
/drivers/infiniband/hw/mlx4/ |
H A D | main.c | 1971 int i, j, eq; local 1981 /* Init eq table */ 1994 eq = 0; 2001 &ibdev->eq_table[eq])) { 2003 pr_warn("Can't allocate EQ %d; reverting to legacy\n", eq); 2004 ibdev->eq_table[eq] = 2005 (eq % dev->caps.num_comp_vectors); 2007 eq++; 2012 for (i = 0, eq = added_eqs; i < dev->caps.num_comp_vectors; i++) 2013 ibdev->eq_table[eq [all...] |
/drivers/net/ethernet/tile/ |
H A D | tilepro.c | 201 lepp_queue_t *eq; member in struct:tile_net_priv 202 /* Protects "eq". */ 526 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, argument 533 unsigned int comp_head = eq->comp_head; 534 unsigned int comp_busy = eq->comp_busy; 537 comps[n++] = eq->comps[comp_head]; 544 eq->comp_head = comp_head; 557 lepp_queue_t *eq = priv->eq; local 567 eq 1736 lepp_queue_t *eq = priv->eq; local 1865 lepp_queue_t *eq = priv->eq; local [all...] |
/drivers/scsi/csiostor/ |
H A D | csio_wr.h | 411 struct csio_eq eq; member in union:csio_q::__anon5298 468 #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) 472 ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) 479 #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
|
/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 94 static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) argument 96 return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe)); 99 static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) argument 101 eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1); 414 static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument 429 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, 434 eq->q.id = rsp->vector_eqid & 0xffff; 435 eq 441 ocrdma_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq, u16 q_len) argument 463 ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument 474 _ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument 482 ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument 504 ocrdma_mbx_mq_cq_create(struct ocrdma_dev *dev, struct ocrdma_queue_info *cq, struct ocrdma_queue_info *eq) argument 906 struct ocrdma_eq *eq = handle; local [all...] |
H A D | ocrdma_verbs.c | 991 struct ocrdma_eq *eq = NULL; local 1001 eq = &dev->eq_tbl[indx]; 1002 irq = ocrdma_get_irq(dev, eq);
|