/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo.h | 84 void nouveau_bo_fence(struct nouveau_bo *, struct nouveau_fence *, bool exclusive);
|
H A D | nouveau_fence.h | 29 int nouveau_fence_sync(struct nouveau_bo *, struct nouveau_channel *, bool exclusive, bool intr);
|
H A D | nouveau_fence.c | 391 nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool exclusive, bool intr) argument 400 if (!exclusive) { 410 if (fence && (!exclusive || !fobj || !fobj->shared_count)) { 429 if (!exclusive || !fobj)
|
H A D | nouveau_bo.c | 1461 nouveau_bo_fence(struct nouveau_bo *nvbo, struct nouveau_fence *fence, bool exclusive) argument 1465 if (exclusive)
|
/drivers/mtd/ubi/ |
H A D | kapi.c | 116 * mode, read-write mode, or exclusive mode. The exclusive mode guarantees that 173 if (vol->exclusive) 179 if (vol->exclusive || vol->writers > 0) 185 if (vol->exclusive || vol->writers || vol->readers) 187 vol->exclusive = 1; 345 vol->exclusive = 0;
|
H A D | cdev.c | 50 * get_exclusive - get exclusive access to an UBI volume. 53 * This function changes UBI volume open mode to "exclusive". Returns previous 63 users = vol->readers + vol->writers + vol->exclusive; 70 vol->exclusive = 1; 80 * revoke_exclusive - revoke exclusive mode. 90 ubi_assert(vol->exclusive == 1 && desc->mode == UBI_EXCLUSIVE); 91 vol->exclusive = 0; 97 vol->exclusive = 1;
|
H A D | ubi.h | 262 * @exclusive: whether somebody holds this volume in exclusive mode 310 int exclusive; member in struct:ubi_volume 391 * @vol->readers, @vol->writers, @vol->exclusive,
|
/drivers/watchdog/ |
H A D | it87_wdt.c | 147 static int exclusive = DEFAULT_EXCLUSIVE; variable 158 module_param(exclusive, int, 0); 159 MODULE_PARM_DESC(exclusive, "Watchdog exclusive device open, default=" 176 * Try to reserve REG and REG + 1 for exclusive access. 399 if (exclusive && test_and_set_bit(WDTS_DEV_OPEN, &wdt_status)) 735 pr_info("Chip IT%04x revision %d initialized. timeout=%d sec (nowayout=%d testmode=%d exclusive=%d nogameport=%d nocir=%d)\n", 737 nowayout, testmode, exclusive, nogameport, nocir);
|
/drivers/staging/lustre/include/linux/libcfs/ |
H A D | libcfs_hash.h | 277 void (*hs_lock)(union cfs_hash_lock *lock, int exclusive); 279 void (*hs_unlock)(union cfs_hash_lock *lock, int exclusive); 281 void (*hs_bkt_lock)(union cfs_hash_lock *lock, int exclusive); 283 void (*hs_bkt_unlock)(union cfs_hash_lock *lock, int exclusive);
|
/drivers/staging/lustre/lustre/libcfs/ |
H A D | hash.c | 122 cfs_hash_nl_lock(union cfs_hash_lock *lock, int exclusive) {} argument 125 cfs_hash_nl_unlock(union cfs_hash_lock *lock, int exclusive) {} argument 128 cfs_hash_spin_lock(union cfs_hash_lock *lock, int exclusive) argument 134 cfs_hash_spin_unlock(union cfs_hash_lock *lock, int exclusive) argument 140 cfs_hash_rw_lock(union cfs_hash_lock *lock, int exclusive) argument 142 if (!exclusive) 149 cfs_hash_rw_unlock(union cfs_hash_lock *lock, int exclusive) argument 151 if (!exclusive)
|
/drivers/target/sbp/ |
H A D | sbp_target.h | 131 int exclusive; member in struct:sbp_login_descriptor
|
H A D | sbp_target.c | 130 int exclusive) 146 if (!exclusive || login->exclusive) 358 * check exclusive bit in login request 363 pr_warn("refusing exclusive login with other active logins\n"); 372 * check exclusive bit in any existing login descriptor 373 * reject with access_denied if any exclusive logins present 376 pr_warn("refusing login while another exclusive login present\n"); 446 login->exclusive = LOGIN_ORB_EXCLUSIVE(be32_to_cpu(req->orb.misc)); 127 sbp_login_count_all_by_lun( struct sbp_tpg *tpg, struct se_lun *lun, int exclusive) argument
|
/drivers/isdn/i4l/ |
H A D | isdn_net.c | 2196 * down. However this channel may be bound exclusive. If the 2199 * We exchange the exclusive bindings of the two channels, the 2209 * exclusive too. */ 2234 /* Now check for exclusive binding again */ 2552 lp->exclusive = -1; 2787 if (cfg->exclusive > 0) { 2790 /* If binding is exclusive, try to grab the channel */ 2796 lp->exclusive = -1; 2804 lp->exclusive = i; 2806 /* Non-exclusive bindin [all...] |
H A D | isdn_ppp.c | 165 char exclusive[ISDN_MAX_CHANNELS]; /* exclusive flags */ local 166 memset(exclusive, 0, ISDN_MAX_CHANNELS); 167 while (net_dev) { /* step through net devices to find exclusive minors */ 170 exclusive[lp->pppbind] = 1; 177 if (ippp_table[i]->state == IPPP_OPEN && !exclusive[ippp_table[i]->minor]) { /* OPEN, but not connected! */
|
/drivers/regulator/ |
H A D | core.c | 1325 bool exclusive, bool allow_dummy) 1374 } else if (!have_full_constraints() || exclusive) { 1382 if (rdev->exclusive) { 1387 if (exclusive && rdev->open_count) { 1403 if (exclusive) { 1404 rdev->exclusive = 1; 1439 * regulator_get_exclusive - obtain exclusive access to a regulator. 1511 rdev->exclusive = 0; 1324 _regulator_get(struct device *dev, const char *id, bool exclusive, bool allow_dummy) argument
|
/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mcg.c | 776 flags |= ctrl->exclusive ? (1 << 2) : 0; 1037 .exclusive = 0, 1348 .exclusive = 0,
|
H A D | en_netdev.c | 171 .exclusive = 1, 517 .exclusive = 0,
|
H A D | en_ethtool.c | 935 .exclusive = 0,
|
/drivers/pci/ |
H A D | pci.c | 2504 * @exclusive: whether the region access is exclusive or not 2511 * If @exclusive is set, then the region is marked so that userspace 2519 const char *res_name, int exclusive) 2533 exclusive)) 2518 __pci_request_region(struct pci_dev *pdev, int bar, const char *res_name, int exclusive) argument
|