[go: nahoru, domu]

Searched refs:eq (Results 1 - 25 of 57) sorted by relevance

123

/drivers/infiniband/hw/ehca/
H A Dehca_eq.c54 struct ehca_eq *eq,
64 spin_lock_init(&eq->spinlock);
65 spin_lock_init(&eq->irq_spinlock);
66 eq->is_initialized = 0;
69 ehca_err(ib_dev, "Invalid EQ type %x. eq=%p", type, eq);
73 ehca_err(ib_dev, "EQ length must not be zero. eq=%p", eq);
78 &eq->pf,
81 &eq
53 ehca_create_eq(struct ehca_shca *shca, struct ehca_eq *eq, const enum ehca_eq_type type, const u32 length) argument
155 ehca_poll_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument
167 ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq) argument
[all...]
H A Dehca_irq.c507 tasklet_hi_schedule(&shca->eq.interrupt_task);
551 struct ehca_eq *eq = &shca->eq; local
552 struct ehca_eqe_cache_entry *eqe_cache = eq->eqe_cache;
557 spin_lock(&eq->irq_spinlock);
564 shca->ipz_hca_handle, eq->ist);
577 eqe_cache[eqe_cnt].eqe = ehca_poll_eq(shca, eq);
604 ret = hipz_h_eoi(eq->ist);
614 if (eq->eqe_cache[i].cq)
615 reset_eq_pending(eq
[all...]
H A Dehca_iverbs.h121 int ehca_create_eq(struct ehca_shca *shca, struct ehca_eq *eq,
124 int ehca_destroy_eq(struct ehca_shca *shca, struct ehca_eq *eq);
126 void *ehca_poll_eq(struct ehca_shca *shca, struct ehca_eq *eq);
H A Dehca_main.c771 ret = ehca_create_eq(shca, &shca->eq, EHCA_EQ, eq_size);
871 ret = ehca_destroy_eq(shca, &shca->eq);
913 ret = ehca_destroy_eq(shca, &shca->eq);
957 if (shca->eq.is_initialized) {
958 /* call deadman proc only if eq ptr does not change */
959 struct ehca_eq *eq = &shca->eq; local
963 spin_lock_irqsave(&eq->spinlock, flags);
964 q_ofs = eq->ipz_queue.current_q_offset;
965 spin_unlock_irqrestore(&eq
[all...]
/drivers/staging/lustre/lnet/lnet/
H A Dlib-eq.c36 * lnet/lnet/lib-eq.c
73 lnet_eq_t *eq; local
93 eq = lnet_eq_alloc();
94 if (eq == NULL)
98 LIBCFS_ALLOC(eq->eq_events, count * sizeof(lnet_event_t));
99 if (eq->eq_events == NULL)
105 eq->eq_deq_seq = 1;
106 eq->eq_enq_seq = 1;
107 eq->eq_size = count;
108 eq
155 struct lnet_eq *eq; local
210 lnet_eq_enqueue_event(lnet_eq_t *eq, lnet_event_t *ev) argument
239 lnet_eq_dequeue_event(lnet_eq_t *eq, lnet_event_t *ev) argument
410 lnet_eq_t *eq = lnet_handle2eq(&eventqs[i]); local
[all...]
H A DMakefile3 lnet-y := api-ni.o config.o lib-me.o lib-msg.o lib-eq.o \
/drivers/net/ethernet/mellanox/mlx5/core/
H A Deq.c105 static struct mlx5_eqe *get_eqe(struct mlx5_eq *eq, u32 entry) argument
107 return mlx5_buf_offset(&eq->buf, entry * MLX5_EQE_SIZE);
110 static struct mlx5_eqe *next_eqe_sw(struct mlx5_eq *eq) argument
112 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1));
114 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe;
186 static void eq_update_ci(struct mlx5_eq *eq, int arm) argument
188 __be32 __iomem *addr = eq
195 mlx5_eq_int(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument
312 struct mlx5_eq *eq = eq_ptr; local
321 init_eq_buf(struct mlx5_eq *eq) argument
332 mlx5_create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, u8 vecidx, int nent, u64 mask, const char *name, struct mlx5_uar *uar) argument
411 mlx5_destroy_unmap_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument
509 mlx5_core_eq_query(struct mlx5_core_dev *dev, struct mlx5_eq *eq, struct mlx5_query_eq_mbox_out *out, int outlen) argument
[all...]
H A DMakefile3 mlx5_core-y := main.o cmd.o debugfs.o fw.o eq.o uar.o pagealloc.o \
H A Ddebugfs.c358 static u64 eq_read_field(struct mlx5_core_dev *dev, struct mlx5_eq *eq, argument
372 err = mlx5_core_eq_query(dev, eq, out, sizeof(*out));
374 mlx5_core_warn(dev, "failed to query eq\n");
562 int mlx5_debug_eq_add(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument
570 &eq->dbg, eq->eqn, eq_fields,
571 ARRAY_SIZE(eq_fields), eq);
573 eq->dbg = NULL;
578 void mlx5_debug_eq_remove(struct mlx5_core_dev *dev, struct mlx5_eq *eq) argument
583 if (eq
[all...]
/drivers/infiniband/hw/mthca/
H A Dmthca_eq.c173 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1),
189 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
194 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8);
199 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
202 arbel_set_eq_ci(dev, eq, ci);
204 tavor_set_eq_ci(dev, eq, ci);
228 static inline struct mthca_eqe *get_eqe(struct mthca_eq *eq, u32 entry) argument
230 unsigned long off = (entry & (eq
234 next_eqe_sw(struct mthca_eq *eq) argument
260 mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) argument
420 struct mthca_eq *eq = eq_ptr; local
454 struct mthca_eq *eq = eq_ptr; local
465 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument
588 mthca_free_eq(struct mthca_dev *dev, struct mthca_eq *eq) argument
[all...]
/drivers/net/ethernet/mellanox/mlx4/
H A Deq.c95 static void eq_set_ci(struct mlx4_eq *eq, int req_not) argument
97 __raw_writel((__force u32) cpu_to_be32((eq->cons_index & 0xffffff) |
99 eq->doorbell);
104 static struct mlx4_eqe *get_eqe(struct mlx4_eq *eq, u32 entry, u8 eqe_factor, argument
107 /* (entry & (eq->nent - 1)) gives us a cyclic array */
108 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size;
116 return eq->page_list[offset / PAGE_SIZE].buf + (offset + (eqe_factor ? MLX4_EQ_ENTRY_SIZE : 0)) % PAGE_SIZE;
119 static struct mlx4_eqe *next_eqe_sw(struct mlx4_eq *eq, u8 eqe_factor, u8 size) argument
121 struct mlx4_eqe *eqe = get_eqe(eq, eq
449 mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq) argument
781 struct mlx4_eq *eq = eq_ptr; local
852 mlx4_get_eq_uar(struct mlx4_dev *dev, struct mlx4_eq *eq) argument
886 mlx4_create_eq(struct mlx4_dev *dev, int nent, u8 intr, struct mlx4_eq *eq) argument
999 mlx4_free_eq(struct mlx4_dev *dev, struct mlx4_eq *eq) argument
[all...]
H A DMakefile3 mlx4_core-y := alloc.o catas.o cmd.o cq.o eq.o fw.o icm.o intf.o main.o mcg.o \
H A Dcq.c277 cq_context->comp_eqn = priv->eq_table.eq[vector].eqn;
296 cq->irq = priv->eq_table.eq[cq->vector].irq;
321 synchronize_irq(priv->eq_table.eq[cq->vector].irq);
/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.c256 struct ehea_eq *eq; local
258 eq = kzalloc(sizeof(*eq), GFP_KERNEL);
259 if (!eq)
262 eq->adapter = adapter;
263 eq->attr.type = type;
264 eq->attr.max_nr_of_eqes = max_nr_of_eqes;
265 eq->attr.eqe_gen = eqe_gen;
266 spin_lock_init(&eq->spinlock);
269 &eq
323 ehea_poll_eq(struct ehea_eq *eq) argument
335 ehea_destroy_eq_res(struct ehea_eq *eq, u64 force) argument
354 ehea_destroy_eq(struct ehea_eq *eq) argument
[all...]
/drivers/net/ethernet/tile/
H A Dtilepro.c201 lepp_queue_t *eq; member in struct:tile_net_priv
202 /* Protects "eq". */
526 static unsigned int tile_net_lepp_grab_comps(lepp_queue_t *eq, argument
533 unsigned int comp_head = eq->comp_head;
534 unsigned int comp_busy = eq->comp_busy;
537 comps[n++] = eq->comps[comp_head];
544 eq->comp_head = comp_head;
557 lepp_queue_t *eq = priv->eq; local
567 eq
1736 lepp_queue_t *eq = priv->eq; local
1865 lepp_queue_t *eq = priv->eq; local
[all...]
/drivers/staging/lustre/include/linux/lnet/
H A Dlib-lnet.h220 lnet_eq_t *eq; local
225 eq = (lnet_eq_t *)lnet_freelist_alloc(&rec->rec_freelist);
228 return eq;
232 lnet_eq_free_locked(lnet_eq_t *eq) argument
238 lnet_freelist_free(&rec->rec_freelist, eq);
242 lnet_eq_free(lnet_eq_t *eq) argument
245 lnet_eq_free_locked(eq);
365 lnet_eq_t *eq; local
367 LIBCFS_ALLOC(eq, sizeof(*eq));
372 lnet_eq_free(lnet_eq_t *eq) argument
478 lnet_eq2handle(lnet_handle_eq_t *handle, lnet_eq_t *eq) argument
[all...]
/drivers/clk/spear/
H A Dspear1340_clock.c268 {.xscale = 5, .yscale = 122, .eq = 0},
270 {.xscale = 10, .yscale = 204, .eq = 0},
272 {.xscale = 4, .yscale = 25, .eq = 0},
274 {.xscale = 4, .yscale = 21, .eq = 0},
276 {.xscale = 5, .yscale = 18, .eq = 0},
278 {.xscale = 2, .yscale = 6, .eq = 0},
280 {.xscale = 5, .yscale = 12, .eq = 0},
282 {.xscale = 2, .yscale = 4, .eq = 0},
284 {.xscale = 5, .yscale = 18, .eq = 1},
286 {.xscale = 1, .yscale = 3, .eq
[all...]
H A Dspear1310_clock.c256 {.xscale = 10, .yscale = 204, .eq = 0}, /* 12.29 MHz */
257 {.xscale = 4, .yscale = 21, .eq = 0}, /* 48 MHz */
258 {.xscale = 2, .yscale = 6, .eq = 0}, /* 83 MHz */
259 {.xscale = 2, .yscale = 4, .eq = 0}, /* 125 MHz */
260 {.xscale = 1, .yscale = 3, .eq = 1}, /* 166 MHz */
261 {.xscale = 1, .yscale = 2, .eq = 1}, /* 250 MHz */
267 {.xscale = 2, .yscale = 6, .eq = 0}, /* divided by 6 */
268 {.xscale = 2, .yscale = 4, .eq = 0}, /* divided by 4 */
269 {.xscale = 1, .yscale = 3, .eq = 1}, /* divided by 3 */
270 {.xscale = 1, .yscale = 2, .eq
[all...]
H A Dclk-aux-synth.c23 * Aux synth gives rate for different values of eq, x and y
49 u8 eq = rtbl[index].eq ? 1 : 2; local
52 (rtbl[index].yscale * eq)) * 10000;
115 val |= (rtbl[i].eq & aux->masks->eq_sel_mask) <<
H A Dspear3xx_clock.c109 {.xscale = 1, .yscale = 81, .eq = 0}, /* 2.049 MHz */
110 {.xscale = 1, .yscale = 59, .eq = 0}, /* 2.822 MHz */
111 {.xscale = 2, .yscale = 81, .eq = 0}, /* 4.098 MHz */
112 {.xscale = 3, .yscale = 89, .eq = 0}, /* 5.644 MHz */
113 {.xscale = 4, .yscale = 81, .eq = 0}, /* 8.197 MHz */
114 {.xscale = 4, .yscale = 59, .eq = 0}, /* 11.254 MHz */
115 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
116 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
117 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
118 {.xscale = 1, .yscale = 2, .eq
[all...]
H A Dspear6xx_clock.c94 {.xscale = 2, .yscale = 27, .eq = 0}, /* 12.296 MHz */
95 {.xscale = 2, .yscale = 8, .eq = 0}, /* 41.5 MHz */
96 {.xscale = 2, .yscale = 4, .eq = 0}, /* 83 MHz */
97 {.xscale = 1, .yscale = 2, .eq = 1}, /* 166 MHz */
H A Dclk.h46 u8 eq; member in struct:aux_rate_tbl
/drivers/infiniband/hw/mlx5/
H A Dmain.c68 struct mlx5_eq *eq, *n; local
72 list_for_each_entry_safe(eq, n, &dev->eqs_list, list) {
73 if (eq->index == vector) {
74 *eqn = eq->eqn;
75 *irqn = eq->irqn;
89 struct mlx5_eq *eq, *n; local
99 eq = kzalloc(sizeof(*eq), GFP_KERNEL);
100 if (!eq) {
106 err = mlx5_create_map_eq(dev->mdev, eq,
140 struct mlx5_eq *eq, *n; local
[all...]
/drivers/scsi/csiostor/
H A Dcsio_wr.h411 struct csio_eq eq; member in union:csio_q::__anon5298
468 #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid)
472 ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid)
479 #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap)
/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.c94 static inline void *ocrdma_get_eqe(struct ocrdma_eq *eq) argument
96 return eq->q.va + (eq->q.tail * sizeof(struct ocrdma_eqe));
99 static inline void ocrdma_eq_inc_tail(struct ocrdma_eq *eq) argument
101 eq->q.tail = (eq->q.tail + 1) & (OCRDMA_EQ_LEN - 1);
414 static int ocrdma_mbx_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument
429 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma,
434 eq->q.id = rsp->vector_eqid & 0xffff;
435 eq
441 ocrdma_create_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq, u16 q_len) argument
463 ocrdma_get_irq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument
474 _ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument
482 ocrdma_destroy_eq(struct ocrdma_dev *dev, struct ocrdma_eq *eq) argument
504 ocrdma_mbx_mq_cq_create(struct ocrdma_dev *dev, struct ocrdma_queue_info *cq, struct ocrdma_queue_info *eq) argument
906 struct ocrdma_eq *eq = handle; local
[all...]

Completed in 405 milliseconds

123