/linux-6.12.1/fs/ext4/ |
D | block_validity.c | 73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local 88 new_entry = kmem_cache_alloc(ext4_system_zone_cachep, in add_system_zone() 90 if (!new_entry) in add_system_zone() 92 new_entry->start_blk = start_blk; in add_system_zone() 93 new_entry->count = count; in add_system_zone() 94 new_entry->ino = ino; in add_system_zone() 95 new_node = &new_entry->node; in add_system_zone() 104 if (can_merge(entry, new_entry)) { in add_system_zone() 105 new_entry->start_blk = entry->start_blk; in add_system_zone() 106 new_entry->count += entry->count; in add_system_zone() [all …]
|
D | mballoc.c | 6285 struct ext4_free_data *new_entry, in ext4_try_merge_freed_extent() argument 6288 if ((entry->efd_tid != new_entry->efd_tid) || in ext4_try_merge_freed_extent() 6289 (entry->efd_group != new_entry->efd_group)) in ext4_try_merge_freed_extent() 6292 new_entry->efd_start_cluster) { in ext4_try_merge_freed_extent() 6293 new_entry->efd_start_cluster = entry->efd_start_cluster; in ext4_try_merge_freed_extent() 6294 new_entry->efd_count += entry->efd_count; in ext4_try_merge_freed_extent() 6295 } else if (new_entry->efd_start_cluster + new_entry->efd_count == in ext4_try_merge_freed_extent() 6297 new_entry->efd_count += entry->efd_count; in ext4_try_merge_freed_extent() 6309 struct ext4_free_data *new_entry) in ext4_mb_free_metadata() argument 6313 ext4_grpblk_t clusters = new_entry->efd_count; in ext4_mb_free_metadata() [all …]
|
/linux-6.12.1/fs/btrfs/ |
D | lru_cache.c | 107 struct btrfs_lru_cache_entry *new_entry, in btrfs_lru_cache_store() argument 110 const u64 key = new_entry->key; in btrfs_lru_cache_store() 121 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store() 126 if (match_entry(head, key, new_entry->gen) != NULL) in btrfs_lru_cache_store() 128 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store() 143 list_add_tail(&new_entry->lru_list, &cache->lru_list); in btrfs_lru_cache_store()
|
D | lru_cache.h | 65 struct btrfs_lru_cache_entry *new_entry,
|
/linux-6.12.1/mm/ |
D | shmem_quota.c | 170 struct quota_id *new_entry, *entry; in shmem_acquire_dquot() local 193 new_entry = kzalloc(sizeof(struct quota_id), GFP_NOFS); in shmem_acquire_dquot() 194 if (!new_entry) { in shmem_acquire_dquot() 199 new_entry->id = id; in shmem_acquire_dquot() 201 new_entry->bhardlimit = sbinfo->qlimits.usrquota_bhardlimit; in shmem_acquire_dquot() 202 new_entry->ihardlimit = sbinfo->qlimits.usrquota_ihardlimit; in shmem_acquire_dquot() 204 new_entry->bhardlimit = sbinfo->qlimits.grpquota_bhardlimit; in shmem_acquire_dquot() 205 new_entry->ihardlimit = sbinfo->qlimits.grpquota_ihardlimit; in shmem_acquire_dquot() 208 new_node = &new_entry->node; in shmem_acquire_dquot() 211 entry = new_entry; in shmem_acquire_dquot()
|
D | mremap.c | 460 void *old_entry, void *new_entry, bool need_rmap_locks) in move_pgt_entry() argument 471 new_entry); in move_pgt_entry() 475 new_entry); in move_pgt_entry() 480 new_entry); in move_pgt_entry() 485 new_entry); in move_pgt_entry()
|
/linux-6.12.1/security/tomoyo/ |
D | domain.c | 31 int tomoyo_update_policy(struct tomoyo_acl_head *new_entry, const int size, in tomoyo_update_policy() argument 48 if (!check_duplicate(entry, new_entry)) in tomoyo_update_policy() 55 entry = tomoyo_commit_ok(new_entry, size); in tomoyo_update_policy() 92 int tomoyo_update_domain(struct tomoyo_acl_info *new_entry, const int size, in tomoyo_update_domain() argument 108 new_entry->cond = tomoyo_get_condition(param); in tomoyo_update_domain() 109 if (!new_entry->cond) in tomoyo_update_domain() 115 if (new_entry->cond->transit && in tomoyo_update_domain() 116 !(new_entry->type == TOMOYO_TYPE_PATH_ACL && in tomoyo_update_domain() 117 container_of(new_entry, struct tomoyo_path_acl, head) in tomoyo_update_domain() 127 if (!tomoyo_same_acl_head(entry, new_entry) || in tomoyo_update_domain() [all …]
|
/linux-6.12.1/drivers/s390/cio/ |
D | qdio_debug.c | 64 struct qdio_dbf_entry *new_entry; in qdio_allocate_dbf() local 85 new_entry = kzalloc(sizeof(struct qdio_dbf_entry), GFP_KERNEL); in qdio_allocate_dbf() 86 if (!new_entry) { in qdio_allocate_dbf() 90 strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); in qdio_allocate_dbf() 91 new_entry->dbf_info = irq_ptr->debug_area; in qdio_allocate_dbf() 93 list_add(&new_entry->dbf_list, &qdio_dbf_list); in qdio_allocate_dbf()
|
/linux-6.12.1/tools/perf/util/ |
D | rblist.c | 13 int rblist__add_node(struct rblist *rblist, const void *new_entry) in rblist__add_node() argument 24 rc = rblist->node_cmp(parent, new_entry); in rblist__add_node() 35 new_node = rblist->node_new(rblist, new_entry); in rblist__add_node()
|
D | rblist.h | 27 struct rb_node *(*node_new)(struct rblist *rlist, const void *new_entry); 34 int rblist__add_node(struct rblist *rblist, const void *new_entry);
|
D | strlist.c | 63 int strlist__add(struct strlist *slist, const char *new_entry) in strlist__add() argument 65 return rblist__add_node(&slist->rblist, new_entry); in strlist__add()
|
/linux-6.12.1/fs/bcachefs/ |
D | replicas.c | 220 struct bch_replicas_entry_v1 *new_entry) in cpu_replicas_add_entry() argument 225 replicas_entry_bytes(new_entry)), in cpu_replicas_add_entry() 238 new_entry, in cpu_replicas_add_entry() 239 replicas_entry_bytes(new_entry)); in cpu_replicas_add_entry() 298 struct bch_replicas_entry_v1 *new_entry) in bch2_mark_replicas_slowpath() argument 303 verify_replicas_entry(new_entry); in bch2_mark_replicas_slowpath() 311 !__replicas_has_entry(&c->replicas_gc, new_entry)) { in bch2_mark_replicas_slowpath() 312 new_gc = cpu_replicas_add_entry(c, &c->replicas_gc, new_entry); in bch2_mark_replicas_slowpath() 319 if (!__replicas_has_entry(&c->replicas, new_entry)) { in bch2_mark_replicas_slowpath() 320 new_r = cpu_replicas_add_entry(c, &c->replicas, new_entry); in bch2_mark_replicas_slowpath()
|
/linux-6.12.1/net/mptcp/ |
D | pm_userspace.c | 126 struct mptcp_pm_addr_entry *entry = NULL, *e, new_entry; in mptcp_userspace_pm_get_local_id() local 141 memset(&new_entry, 0, sizeof(struct mptcp_pm_addr_entry)); in mptcp_userspace_pm_get_local_id() 142 new_entry.addr = *skc; in mptcp_userspace_pm_get_local_id() 143 new_entry.addr.id = 0; in mptcp_userspace_pm_get_local_id() 144 new_entry.flags = MPTCP_PM_ADDR_FLAG_IMPLICIT; in mptcp_userspace_pm_get_local_id() 146 if (new_entry.addr.port == msk_sport) in mptcp_userspace_pm_get_local_id() 147 new_entry.addr.port = 0; in mptcp_userspace_pm_get_local_id() 149 return mptcp_userspace_pm_append_new_local_addr(msk, &new_entry, true); in mptcp_userspace_pm_get_local_id()
|
/linux-6.12.1/arch/x86/kvm/svm/ |
D | avic.c | 279 u64 *entry, new_entry; in avic_init_backing_page() local 311 new_entry = __sme_set((page_to_phys(svm->avic_backing_page) & in avic_init_backing_page() 314 WRITE_ONCE(*entry, new_entry); in avic_init_backing_page() 581 u32 *entry, new_entry; in avic_ldr_write() local 588 new_entry = READ_ONCE(*entry); in avic_ldr_write() 589 new_entry &= ~AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK; in avic_ldr_write() 590 new_entry |= (g_physical_id & AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK); in avic_ldr_write() 591 new_entry |= AVIC_LOGICAL_ID_ENTRY_VALID_MASK; in avic_ldr_write() 592 WRITE_ONCE(*entry, new_entry); in avic_ldr_write()
|
/linux-6.12.1/fs/smb/client/ |
D | readdir.c | 447 char *new_entry; in nxt_dir_entry() local 454 new_entry = old_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 + in nxt_dir_entry() 463 new_entry = old_entry + next_offset; in nxt_dir_entry() 465 cifs_dbg(FYI, "new entry %p old entry %p\n", new_entry, old_entry); in nxt_dir_entry() 467 if (new_entry >= end_of_smb) { in nxt_dir_entry() 469 new_entry, end_of_smb, old_entry); in nxt_dir_entry() 472 (new_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 > end_of_smb)) in nxt_dir_entry() 474 (new_entry + sizeof(FILE_DIRECTORY_INFO) + 1 > end_of_smb))) { in nxt_dir_entry() 476 new_entry, end_of_smb); in nxt_dir_entry() 479 return new_entry; in nxt_dir_entry()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/ |
D | mcg.c | 155 struct mlx4_steer_index *new_entry; in new_steering_entry() local 165 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in new_steering_entry() 166 if (!new_entry) in new_steering_entry() 169 INIT_LIST_HEAD(&new_entry->duplicates); in new_steering_entry() 170 new_entry->index = index; in new_steering_entry() 171 list_add_tail(&new_entry->list, &s_steer->steer_entries[steer]); in new_steering_entry() 184 list_add_tail(&dqp->list, &new_entry->duplicates); in new_steering_entry() 233 list_del(&new_entry->list); in new_steering_entry() 234 kfree(new_entry); in new_steering_entry() 1117 u8 new_entry = 0; in mlx4_qp_attach_common() local [all …]
|
/linux-6.12.1/fs/f2fs/ |
D | namei.c | 897 struct f2fs_dir_entry *new_entry; in f2fs_rename() local 976 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, in f2fs_rename() 978 if (!new_entry) { in f2fs_rename() 992 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_rename() 1090 struct f2fs_dir_entry *old_entry, *new_entry; in f2fs_cross_rename() local 1123 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_page); in f2fs_cross_rename() 1124 if (!new_entry) { in f2fs_cross_rename() 1200 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_cross_rename()
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | affinity.c | 596 bool new_entry = false; in hfi1_dev_affinity_init() local 617 new_entry = true; in hfi1_dev_affinity_init() 694 ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); in hfi1_dev_affinity_init() 698 if (new_entry) in hfi1_dev_affinity_init() 707 if (new_entry) in hfi1_dev_affinity_init()
|
/linux-6.12.1/drivers/scsi/ |
D | nsp32.c | 1723 int new_entry; in nsp32_adjust_busfree() local 1739 for (new_entry = old_entry; new_entry < sg_num; new_entry++) { in nsp32_adjust_busfree() 1740 sentlen += (le32_to_cpu(sgt[new_entry].len) & ~SGTEND); in nsp32_adjust_busfree() 1747 if (new_entry == sg_num) { in nsp32_adjust_busfree() 1763 len = le32_to_cpu(sgt[new_entry].len); in nsp32_adjust_busfree() 1764 addr = le32_to_cpu(sgt[new_entry].addr); in nsp32_adjust_busfree() 1766 sgt[new_entry].addr = cpu_to_le32(addr); in nsp32_adjust_busfree() 1767 sgt[new_entry].len = cpu_to_le32(restlen); in nsp32_adjust_busfree() 1770 data->cur_entry = new_entry; in nsp32_adjust_busfree()
|
D | hpsa.c | 1363 int entry, struct hpsa_scsi_dev_t *new_entry) in hpsa_scsi_update_entry() argument 1369 h->dev[entry]->raid_level = new_entry->raid_level; in hpsa_scsi_update_entry() 1374 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1377 if (new_entry->offload_config && new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1386 h->dev[entry]->raid_map = new_entry->raid_map; in hpsa_scsi_update_entry() 1387 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1389 if (new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry() 1390 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry() 1393 h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled; in hpsa_scsi_update_entry() 1394 h->dev[entry]->offload_config = new_entry->offload_config; in hpsa_scsi_update_entry() [all …]
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | cxgb4vf_main.c | 275 struct hash_mac_addr *new_entry, *entry; in cxgb4vf_change_mac() local 291 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in cxgb4vf_change_mac() 292 if (!new_entry) in cxgb4vf_change_mac() 294 ether_addr_copy(new_entry->addr, addr); in cxgb4vf_change_mac() 295 new_entry->iface_mac = true; in cxgb4vf_change_mac() 296 list_add_tail(&new_entry->list, &adapter->mac_hlist); in cxgb4vf_change_mac() 945 struct hash_mac_addr *new_entry; in cxgb4vf_mac_sync() local 956 new_entry = kzalloc(sizeof(*new_entry), GFP_ATOMIC); in cxgb4vf_mac_sync() 957 if (!new_entry) in cxgb4vf_mac_sync() 959 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4vf_mac_sync() [all …]
|
/linux-6.12.1/drivers/iommu/ |
D | virtio-iommu.c | 944 struct iommu_resv_region *entry, *new_entry, *msi = NULL; in viommu_get_resv_regions() local 952 new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); in viommu_get_resv_regions() 953 if (!new_entry) in viommu_get_resv_regions() 955 list_add_tail(&new_entry->list, head); in viommu_get_resv_regions()
|
/linux-6.12.1/tools/net/ynl/lib/ |
D | nlspec.py | 123 e = self.new_entry(entry, prev_entry, value_start) 128 def new_entry(self, entry, prev_entry, value_start): member in SpecEnumSet
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_acl_tcam.c | 1193 struct mlxsw_sp_acl_tcam_entry *new_entry; in mlxsw_sp_acl_tcam_ventry_migrate() local 1202 new_entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, chunk); in mlxsw_sp_acl_tcam_ventry_migrate() 1203 if (IS_ERR(new_entry)) in mlxsw_sp_acl_tcam_ventry_migrate() 1204 return PTR_ERR(new_entry); in mlxsw_sp_acl_tcam_ventry_migrate() 1206 ventry->entry = new_entry; in mlxsw_sp_acl_tcam_ventry_migrate()
|
/linux-6.12.1/fs/exfat/ |
D | namei.c | 1130 int new_entry = 0; in __exfat_rename() local 1158 new_entry = new_ei->entry; in __exfat_rename() 1159 ep = exfat_get_dentry(sb, p_dir, new_entry, &new_bh); in __exfat_rename() 1201 ret = exfat_get_dentry_set(&es, sb, p_dir, new_entry, in __exfat_rename()
|