Home
last modified time | relevance | path

Searched refs:new_entry (Results 1 – 25 of 41) sorted by relevance

12

/linux-6.12.1/fs/ext4/
Dblock_validity.c73 struct ext4_system_zone *new_entry, *entry; in add_system_zone() local
88 new_entry = kmem_cache_alloc(ext4_system_zone_cachep, in add_system_zone()
90 if (!new_entry) in add_system_zone()
92 new_entry->start_blk = start_blk; in add_system_zone()
93 new_entry->count = count; in add_system_zone()
94 new_entry->ino = ino; in add_system_zone()
95 new_node = &new_entry->node; in add_system_zone()
104 if (can_merge(entry, new_entry)) { in add_system_zone()
105 new_entry->start_blk = entry->start_blk; in add_system_zone()
106 new_entry->count += entry->count; in add_system_zone()
[all …]
Dmballoc.c6285 struct ext4_free_data *new_entry, in ext4_try_merge_freed_extent() argument
6288 if ((entry->efd_tid != new_entry->efd_tid) || in ext4_try_merge_freed_extent()
6289 (entry->efd_group != new_entry->efd_group)) in ext4_try_merge_freed_extent()
6292 new_entry->efd_start_cluster) { in ext4_try_merge_freed_extent()
6293 new_entry->efd_start_cluster = entry->efd_start_cluster; in ext4_try_merge_freed_extent()
6294 new_entry->efd_count += entry->efd_count; in ext4_try_merge_freed_extent()
6295 } else if (new_entry->efd_start_cluster + new_entry->efd_count == in ext4_try_merge_freed_extent()
6297 new_entry->efd_count += entry->efd_count; in ext4_try_merge_freed_extent()
6309 struct ext4_free_data *new_entry) in ext4_mb_free_metadata() argument
6313 ext4_grpblk_t clusters = new_entry->efd_count; in ext4_mb_free_metadata()
[all …]
/linux-6.12.1/fs/btrfs/
Dlru_cache.c107 struct btrfs_lru_cache_entry *new_entry, in btrfs_lru_cache_store() argument
110 const u64 key = new_entry->key; in btrfs_lru_cache_store()
121 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store()
126 if (match_entry(head, key, new_entry->gen) != NULL) in btrfs_lru_cache_store()
128 list_add_tail(&new_entry->list, head); in btrfs_lru_cache_store()
143 list_add_tail(&new_entry->lru_list, &cache->lru_list); in btrfs_lru_cache_store()
Dlru_cache.h65 struct btrfs_lru_cache_entry *new_entry,
/linux-6.12.1/mm/
Dshmem_quota.c170 struct quota_id *new_entry, *entry; in shmem_acquire_dquot() local
193 new_entry = kzalloc(sizeof(struct quota_id), GFP_NOFS); in shmem_acquire_dquot()
194 if (!new_entry) { in shmem_acquire_dquot()
199 new_entry->id = id; in shmem_acquire_dquot()
201 new_entry->bhardlimit = sbinfo->qlimits.usrquota_bhardlimit; in shmem_acquire_dquot()
202 new_entry->ihardlimit = sbinfo->qlimits.usrquota_ihardlimit; in shmem_acquire_dquot()
204 new_entry->bhardlimit = sbinfo->qlimits.grpquota_bhardlimit; in shmem_acquire_dquot()
205 new_entry->ihardlimit = sbinfo->qlimits.grpquota_ihardlimit; in shmem_acquire_dquot()
208 new_node = &new_entry->node; in shmem_acquire_dquot()
211 entry = new_entry; in shmem_acquire_dquot()
Dmremap.c460 void *old_entry, void *new_entry, bool need_rmap_locks) in move_pgt_entry() argument
471 new_entry); in move_pgt_entry()
475 new_entry); in move_pgt_entry()
480 new_entry); in move_pgt_entry()
485 new_entry); in move_pgt_entry()
/linux-6.12.1/security/tomoyo/
Ddomain.c31 int tomoyo_update_policy(struct tomoyo_acl_head *new_entry, const int size, in tomoyo_update_policy() argument
48 if (!check_duplicate(entry, new_entry)) in tomoyo_update_policy()
55 entry = tomoyo_commit_ok(new_entry, size); in tomoyo_update_policy()
92 int tomoyo_update_domain(struct tomoyo_acl_info *new_entry, const int size, in tomoyo_update_domain() argument
108 new_entry->cond = tomoyo_get_condition(param); in tomoyo_update_domain()
109 if (!new_entry->cond) in tomoyo_update_domain()
115 if (new_entry->cond->transit && in tomoyo_update_domain()
116 !(new_entry->type == TOMOYO_TYPE_PATH_ACL && in tomoyo_update_domain()
117 container_of(new_entry, struct tomoyo_path_acl, head) in tomoyo_update_domain()
127 if (!tomoyo_same_acl_head(entry, new_entry) || in tomoyo_update_domain()
[all …]
/linux-6.12.1/drivers/s390/cio/
Dqdio_debug.c64 struct qdio_dbf_entry *new_entry; in qdio_allocate_dbf() local
85 new_entry = kzalloc(sizeof(struct qdio_dbf_entry), GFP_KERNEL); in qdio_allocate_dbf()
86 if (!new_entry) { in qdio_allocate_dbf()
90 strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); in qdio_allocate_dbf()
91 new_entry->dbf_info = irq_ptr->debug_area; in qdio_allocate_dbf()
93 list_add(&new_entry->dbf_list, &qdio_dbf_list); in qdio_allocate_dbf()
/linux-6.12.1/tools/perf/util/
Drblist.c13 int rblist__add_node(struct rblist *rblist, const void *new_entry) in rblist__add_node() argument
24 rc = rblist->node_cmp(parent, new_entry); in rblist__add_node()
35 new_node = rblist->node_new(rblist, new_entry); in rblist__add_node()
Drblist.h27 struct rb_node *(*node_new)(struct rblist *rlist, const void *new_entry);
34 int rblist__add_node(struct rblist *rblist, const void *new_entry);
Dstrlist.c63 int strlist__add(struct strlist *slist, const char *new_entry) in strlist__add() argument
65 return rblist__add_node(&slist->rblist, new_entry); in strlist__add()
/linux-6.12.1/fs/bcachefs/
Dreplicas.c220 struct bch_replicas_entry_v1 *new_entry) in cpu_replicas_add_entry() argument
225 replicas_entry_bytes(new_entry)), in cpu_replicas_add_entry()
238 new_entry, in cpu_replicas_add_entry()
239 replicas_entry_bytes(new_entry)); in cpu_replicas_add_entry()
298 struct bch_replicas_entry_v1 *new_entry) in bch2_mark_replicas_slowpath() argument
303 verify_replicas_entry(new_entry); in bch2_mark_replicas_slowpath()
311 !__replicas_has_entry(&c->replicas_gc, new_entry)) { in bch2_mark_replicas_slowpath()
312 new_gc = cpu_replicas_add_entry(c, &c->replicas_gc, new_entry); in bch2_mark_replicas_slowpath()
319 if (!__replicas_has_entry(&c->replicas, new_entry)) { in bch2_mark_replicas_slowpath()
320 new_r = cpu_replicas_add_entry(c, &c->replicas, new_entry); in bch2_mark_replicas_slowpath()
/linux-6.12.1/net/mptcp/
Dpm_userspace.c126 struct mptcp_pm_addr_entry *entry = NULL, *e, new_entry; in mptcp_userspace_pm_get_local_id() local
141 memset(&new_entry, 0, sizeof(struct mptcp_pm_addr_entry)); in mptcp_userspace_pm_get_local_id()
142 new_entry.addr = *skc; in mptcp_userspace_pm_get_local_id()
143 new_entry.addr.id = 0; in mptcp_userspace_pm_get_local_id()
144 new_entry.flags = MPTCP_PM_ADDR_FLAG_IMPLICIT; in mptcp_userspace_pm_get_local_id()
146 if (new_entry.addr.port == msk_sport) in mptcp_userspace_pm_get_local_id()
147 new_entry.addr.port = 0; in mptcp_userspace_pm_get_local_id()
149 return mptcp_userspace_pm_append_new_local_addr(msk, &new_entry, true); in mptcp_userspace_pm_get_local_id()
/linux-6.12.1/arch/x86/kvm/svm/
Davic.c279 u64 *entry, new_entry; in avic_init_backing_page() local
311 new_entry = __sme_set((page_to_phys(svm->avic_backing_page) & in avic_init_backing_page()
314 WRITE_ONCE(*entry, new_entry); in avic_init_backing_page()
581 u32 *entry, new_entry; in avic_ldr_write() local
588 new_entry = READ_ONCE(*entry); in avic_ldr_write()
589 new_entry &= ~AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK; in avic_ldr_write()
590 new_entry |= (g_physical_id & AVIC_LOGICAL_ID_ENTRY_GUEST_PHYSICAL_ID_MASK); in avic_ldr_write()
591 new_entry |= AVIC_LOGICAL_ID_ENTRY_VALID_MASK; in avic_ldr_write()
592 WRITE_ONCE(*entry, new_entry); in avic_ldr_write()
/linux-6.12.1/fs/smb/client/
Dreaddir.c447 char *new_entry; in nxt_dir_entry() local
454 new_entry = old_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 + in nxt_dir_entry()
463 new_entry = old_entry + next_offset; in nxt_dir_entry()
465 cifs_dbg(FYI, "new entry %p old entry %p\n", new_entry, old_entry); in nxt_dir_entry()
467 if (new_entry >= end_of_smb) { in nxt_dir_entry()
469 new_entry, end_of_smb, old_entry); in nxt_dir_entry()
472 (new_entry + sizeof(FIND_FILE_STANDARD_INFO) + 1 > end_of_smb)) in nxt_dir_entry()
474 (new_entry + sizeof(FILE_DIRECTORY_INFO) + 1 > end_of_smb))) { in nxt_dir_entry()
476 new_entry, end_of_smb); in nxt_dir_entry()
479 return new_entry; in nxt_dir_entry()
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/
Dmcg.c155 struct mlx4_steer_index *new_entry; in new_steering_entry() local
165 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in new_steering_entry()
166 if (!new_entry) in new_steering_entry()
169 INIT_LIST_HEAD(&new_entry->duplicates); in new_steering_entry()
170 new_entry->index = index; in new_steering_entry()
171 list_add_tail(&new_entry->list, &s_steer->steer_entries[steer]); in new_steering_entry()
184 list_add_tail(&dqp->list, &new_entry->duplicates); in new_steering_entry()
233 list_del(&new_entry->list); in new_steering_entry()
234 kfree(new_entry); in new_steering_entry()
1117 u8 new_entry = 0; in mlx4_qp_attach_common() local
[all …]
/linux-6.12.1/fs/f2fs/
Dnamei.c897 struct f2fs_dir_entry *new_entry; in f2fs_rename() local
976 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, in f2fs_rename()
978 if (!new_entry) { in f2fs_rename()
992 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_rename()
1090 struct f2fs_dir_entry *old_entry, *new_entry; in f2fs_cross_rename() local
1123 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_page); in f2fs_cross_rename()
1124 if (!new_entry) { in f2fs_cross_rename()
1200 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_cross_rename()
/linux-6.12.1/drivers/infiniband/hw/hfi1/
Daffinity.c596 bool new_entry = false; in hfi1_dev_affinity_init() local
617 new_entry = true; in hfi1_dev_affinity_init()
694 ret = _dev_comp_vect_cpu_mask_init(dd, entry, new_entry); in hfi1_dev_affinity_init()
698 if (new_entry) in hfi1_dev_affinity_init()
707 if (new_entry) in hfi1_dev_affinity_init()
/linux-6.12.1/drivers/scsi/
Dnsp32.c1723 int new_entry; in nsp32_adjust_busfree() local
1739 for (new_entry = old_entry; new_entry < sg_num; new_entry++) { in nsp32_adjust_busfree()
1740 sentlen += (le32_to_cpu(sgt[new_entry].len) & ~SGTEND); in nsp32_adjust_busfree()
1747 if (new_entry == sg_num) { in nsp32_adjust_busfree()
1763 len = le32_to_cpu(sgt[new_entry].len); in nsp32_adjust_busfree()
1764 addr = le32_to_cpu(sgt[new_entry].addr); in nsp32_adjust_busfree()
1766 sgt[new_entry].addr = cpu_to_le32(addr); in nsp32_adjust_busfree()
1767 sgt[new_entry].len = cpu_to_le32(restlen); in nsp32_adjust_busfree()
1770 data->cur_entry = new_entry; in nsp32_adjust_busfree()
Dhpsa.c1363 int entry, struct hpsa_scsi_dev_t *new_entry) in hpsa_scsi_update_entry() argument
1369 h->dev[entry]->raid_level = new_entry->raid_level; in hpsa_scsi_update_entry()
1374 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1377 if (new_entry->offload_config && new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry()
1386 h->dev[entry]->raid_map = new_entry->raid_map; in hpsa_scsi_update_entry()
1387 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1389 if (new_entry->offload_to_be_enabled) { in hpsa_scsi_update_entry()
1390 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle; in hpsa_scsi_update_entry()
1393 h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled; in hpsa_scsi_update_entry()
1394 h->dev[entry]->offload_config = new_entry->offload_config; in hpsa_scsi_update_entry()
[all …]
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/
Dcxgb4vf_main.c275 struct hash_mac_addr *new_entry, *entry; in cxgb4vf_change_mac() local
291 new_entry = kzalloc(sizeof(*new_entry), GFP_KERNEL); in cxgb4vf_change_mac()
292 if (!new_entry) in cxgb4vf_change_mac()
294 ether_addr_copy(new_entry->addr, addr); in cxgb4vf_change_mac()
295 new_entry->iface_mac = true; in cxgb4vf_change_mac()
296 list_add_tail(&new_entry->list, &adapter->mac_hlist); in cxgb4vf_change_mac()
945 struct hash_mac_addr *new_entry; in cxgb4vf_mac_sync() local
956 new_entry = kzalloc(sizeof(*new_entry), GFP_ATOMIC); in cxgb4vf_mac_sync()
957 if (!new_entry) in cxgb4vf_mac_sync()
959 ether_addr_copy(new_entry->addr, mac_addr); in cxgb4vf_mac_sync()
[all …]
/linux-6.12.1/drivers/iommu/
Dvirtio-iommu.c944 struct iommu_resv_region *entry, *new_entry, *msi = NULL; in viommu_get_resv_regions() local
952 new_entry = kmemdup(entry, sizeof(*entry), GFP_KERNEL); in viommu_get_resv_regions()
953 if (!new_entry) in viommu_get_resv_regions()
955 list_add_tail(&new_entry->list, head); in viommu_get_resv_regions()
/linux-6.12.1/tools/net/ynl/lib/
Dnlspec.py123 e = self.new_entry(entry, prev_entry, value_start)
128 def new_entry(self, entry, prev_entry, value_start): member in SpecEnumSet
/linux-6.12.1/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_acl_tcam.c1193 struct mlxsw_sp_acl_tcam_entry *new_entry; in mlxsw_sp_acl_tcam_ventry_migrate() local
1202 new_entry = mlxsw_sp_acl_tcam_entry_create(mlxsw_sp, ventry, chunk); in mlxsw_sp_acl_tcam_ventry_migrate()
1203 if (IS_ERR(new_entry)) in mlxsw_sp_acl_tcam_ventry_migrate()
1204 return PTR_ERR(new_entry); in mlxsw_sp_acl_tcam_ventry_migrate()
1206 ventry->entry = new_entry; in mlxsw_sp_acl_tcam_ventry_migrate()
/linux-6.12.1/fs/exfat/
Dnamei.c1130 int new_entry = 0; in __exfat_rename() local
1158 new_entry = new_ei->entry; in __exfat_rename()
1159 ep = exfat_get_dentry(sb, p_dir, new_entry, &new_bh); in __exfat_rename()
1201 ret = exfat_get_dentry_set(&es, sb, p_dir, new_entry, in __exfat_rename()

12