/linux-6.12.1/include/dt-bindings/memory/ |
D | tegra194-mc.h | 4 /* special clients */ 8 /* host1x clients */ 25 /* GPU clients */ 28 /* other SoC clients */ 45 /* GPC DMA clients */ 55 /* APE DMA clients */ 65 /* APE DMA clients */ 147 /* MSS internal memqual MIU7 read clients */ 149 /* MSS internal memqual MIU7 write clients */ 151 /* High-definition audio (HDA) read clients */ [all …]
|
D | tegra234-mc.h | 7 /* special clients */ 180 /* MSS internal memqual MIU7 read clients */ 182 /* MSS internal memqual MIU7 write clients */ 184 /* MSS internal memqual MIU8 read clients */ 186 /* MSS internal memqual MIU8 write clients */ 188 /* MSS internal memqual MIU9 read clients */ 190 /* MSS internal memqual MIU9 write clients */ 192 /* MSS internal memqual MIU10 read clients */ 194 /* MSS internal memqual MIU10 write clients */ 196 /* MSS internal memqual MIU11 read clients */ [all …]
|
D | tegra186-mc.h | 4 /* special clients */ 8 /* host1x clients */ 25 /* GPU clients */ 28 /* other SoC clients */ 45 /* GPC DMA clients */ 55 /* APE DMA clients */ 65 /* APE DMA clients */
|
/linux-6.12.1/fs/nfsd/ |
D | Kconfig | 30 available to clients mounting the NFS server on this system. 42 Unless you are hosting ancient (1990's era) NFS clients, you don't 57 This protocol extension allows applications on NFS clients to 63 NFSv3 ACL protocol extension allowing NFS clients to manipulate 65 clients which support the Solaris NFSv3 ACL protocol can then 104 clients to directly perform I/O to block devices accessible to both 105 the server and the clients. See RFC 5663 for more details. 117 clients to directly perform I/O to SCSI devices accessible to both 118 the server and the clients. See draft-ietf-nfsv4-scsi-layout for 130 enables NFS clients to directly perform I/O to NFSv3 devices [all …]
|
/linux-6.12.1/drivers/hsi/ |
D | hsi_boardinfo.c | 3 * HSI clients registration interface 22 * hsi_register_board_info - Register HSI clients information 23 * @info: Array of HSI clients on the board 26 * HSI clients are statically declared and registered on board files. 28 * HSI clients will be automatically registered to the HSI bus once the 29 * controller and the port where the clients wishes to attach are registered
|
/linux-6.12.1/drivers/net/ethernet/qlogic/qed/ |
D | qed_cxt.c | 284 struct qed_ilt_client_cfg *p_cli = &p_mngr->clients[ilt_client]; in qed_cxt_get_ilt_page_size() 319 u32 page_sz = p_mgr->clients[ILT_CLI_CDUC].p_size.val; in qed_cxt_set_proto_cid_count() 416 p_cli = &p_hwfn->p_cxt_mngr->clients[ILT_CLI_CDUC]; in qed_ilt_get_dynamic_line_cnt() 444 struct qed_ilt_client_cfg *clients = p_hwfn->p_cxt_mngr->clients; in qed_cxt_ilt_blk_reset() local 449 clients[cli_idx].pf_blks[blk_idx].total_size = 0; in qed_cxt_ilt_blk_reset() 452 clients[cli_idx].vf_blks[blk_idx].total_size = 0; in qed_cxt_ilt_blk_reset() 485 p_cli = qed_cxt_set_cli(&p_mngr->clients[ILT_CLI_CDUC]); in qed_cxt_cfg_ilt_compute() 492 /* get the counters for the CDUC and QM clients */ in qed_cxt_cfg_ilt_compute() 523 p_cli = qed_cxt_set_cli(&p_mngr->clients[ILT_CLI_CDUT]); in qed_cxt_cfg_ilt_compute() 631 p_cli = qed_cxt_set_cli(&p_mngr->clients[ILT_CLI_QM]); in qed_cxt_cfg_ilt_compute() [all …]
|
/linux-6.12.1/drivers/gpu/vga/ |
D | vga_switcheroo.c | 70 * clients. The mux is called the handler. Muxless machines also register a 76 * there can thus be up to three clients: Two vga clients (GPUs) and one audio 110 * For audio clients, the @fb_info and @active members are bogus. For GPU 111 * clients, the @vga_dev is bogus. 138 * (counting only vga clients, not audio clients) 139 * @clients: list of registered clients 157 struct list_head clients; member 175 .clients = LIST_HEAD_INIT(vgasr_priv.clients), 181 /* we're ready if we get two clients + handler */ in vga_switcheroo_ready() 195 list_for_each_entry(client, &vgasr_priv.clients, list) { in vga_switcheroo_enable() [all …]
|
/linux-6.12.1/Documentation/filesystems/nfs/ |
D | client-identifier.rst | 41 leases. When the server restarts and clients attempt to recover 43 clients that held state before the server restarted and clients 92 circumstances where clients can interfere with each other. This is 95 If distinct clients present the same "co_ownerid" string and use 97 unable to tell that the clients are not the same. Each distinct 102 If distinct clients present the same "co_ownerid" string and use 104 to operate normally but reject subsequent clients with the same 117 to other clients in the meantime. This is referred to as "lock 134 - NFS-root (diskless) clients, where the local DHCP server (or 142 - Clients across multiple administrative domains that access a
|
/linux-6.12.1/include/uapi/linux/ |
D | um_timetravel.h | 202 * Since this is shared memory with all clients and controller and controller 208 * messages (for clients that are not using shared memory, the controller will 209 * handle an offset and make the clients think they start at time zero.) 227 * the um_timetravel_schedshm clients array, that holds info related to the client 253 * set by controller once at init, clients must check this after mapping 256 * @len: Length of all the memory including header (@hdr), clients should once 259 * clients be dynamic based on controller support. 264 * is made by any client. Clients also must update this value when they 272 * @max_clients: size of @clients array, set once at init by the controller. 273 * @clients: clients array see &union um_timetravel_schedshm_client for doc, [all …]
|
/linux-6.12.1/drivers/gpu/host1x/ |
D | bus.c | 19 static LIST_HEAD(clients); 123 list_move_tail(&client->list, &device->clients); in host1x_subdev_register() 157 * it from list of clients. in __host1x_subdev_unregister() 188 * its &host1x_driver.probe implementation to initialize each of its clients. 200 list_for_each_entry(client, &device->clients, list) { in host1x_device_init() 211 list_for_each_entry(client, &device->clients, list) { in host1x_device_init() 228 list_for_each_entry_continue_reverse(client, &device->clients, list) in host1x_device_init() 233 client = list_entry(&device->clients, struct host1x_client, list); in host1x_device_init() 236 list_for_each_entry_continue_reverse(client, &device->clients, list) in host1x_device_init() 250 * function to tear down each of its clients. Typically this is done after a [all …]
|
/linux-6.12.1/include/linux/ |
D | vga_switcheroo.h | 41 * DDC lines separately. This signals to clients that they should call 44 * the AUX channel separately. This signals to clients that the active 77 * @VGA_SWITCHEROO_UNKNOWN_ID: initial identifier assigned to vga clients. 84 * Client identifier. Audio clients use the same identifier & 0x100. 97 * two vga clients have registered. It allows the handler to perform 99 * vga clients. Currently only the radeon and amdgpu drivers use this. 140 * set to NULL. For audio clients, the @reprobe member is bogus. 141 * OTOH, @gpu_bound is only for audio clients, and not used for GPU clients.
|
D | pci-p2pdma.h | 22 int pci_p2pdma_distance_many(struct pci_dev *provider, struct device **clients, 25 struct pci_dev *pci_p2pmem_find_many(struct device **clients, int num_clients); 44 struct device **clients, int num_clients, bool verbose) in pci_p2pdma_distance_many() argument 52 static inline struct pci_dev *pci_p2pmem_find_many(struct device **clients, in pci_p2pmem_find_many() argument
|
/linux-6.12.1/Documentation/gpu/ |
D | tegra.rst | 7 buffer provided directly by the CPU, to its clients via channels. Software, 30 The various host1x clients need to be bound together into a logical device in 35 tree for matching device nodes, adding the required clients to a list. Drivers 36 for individual clients register with the infrastructure as well and are added 39 Once all clients are available, the infrastructure will initialize the logical 41 the subsystem and in turn initialize each of its clients. 43 Similarly, when one of the clients is unregistered, the infrastructure will 45 the subsystem specific bits are torn down and the clients destroyed in turn.
|
D | drm-uapi.rst | 137 clients no longer require running compositors or graphics servers to 138 make use of a GPU. But the DRM API required unprivileged clients to 140 step and to grant clients GPU access without authenticating, render 141 nodes were introduced. Render nodes solely serve render clients, that 146 clients together with the legacy drmAuth authentication procedure. 154 nodes are designed to avoid the buffer-leaks, which occur if clients 158 clients can use them. Driver authors must be careful not to allow any 163 authenticates clients on the privileged primary/legacy node is no longer 165 granted GPU access. Communication between clients (or servers) is done 167 clients must not use the insecure FLINK interface. [all …]
|
/linux-6.12.1/Documentation/userspace-api/ |
D | tee.rst | 30 There are two classes of clients, normal clients and supplicants. The latter is 35 Much of the communication between clients and the TEE is opaque to the 37 clients, forward them to the TEE and send back the results. In the case of
|
/linux-6.12.1/include/net/ |
D | bond_alb.h | 33 #define TLB_HASH_TABLE_SIZE 256 /* The size of the clients hash table. 80 * connection. This is the Clients Hash Table entry struct. 86 * all the clients without walking over all the unused elements of the table. 116 u32 head; /* Index to the head of the bi-directional clients 121 u32 load; /* Each slave sums the loadHistory of all clients 136 * to all rx clients
|
/linux-6.12.1/Documentation/virt/acrn/ |
D | io-request.rst | 23 2. I/O clients 27 GPA falls in a certain range. Multiple I/O clients can be associated with each 30 any other clients. The ACRN userspace acts as the default client for each User 34 I/O requests and I/O clients. 54 || | | | I/O clients | | | | 93 registered clients based on the address of the I/O accesses, updates
|
/linux-6.12.1/Documentation/devicetree/bindings/cache/ |
D | qcom,llcc.yaml | 14 that can be shared by multiple clients. Clients here are different cores in the 15 SoC, the idea is to minimize the local caches at the clients and migrate to 17 which are assigned to clients. Clients can query the slice details, activate
|
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/ |
D | migrate_reuseport.c | 52 __s64 clients[NR_CLIENTS]; member 284 test_case->clients[i] = socket(test_case->family, SOCK_STREAM, in start_clients() 286 if (!ASSERT_NEQ(test_case->clients[i], -1, "socket")) in start_clients() 290 * clients will transition to TCP_ESTABLISHED immediately. in start_clients() 292 err = settimeo(test_case->clients[i], 100); in start_clients() 299 err = setsockopt(test_case->clients[i], IPPROTO_TCP, in start_clients() 307 err = connect(test_case->clients[i], in start_clients() 313 err = write(test_case->clients[i], buf, MSGLEN); in start_clients() 468 init_fds(test_case->clients, NR_CLIENTS); in run_test() 528 close_fds(test_case->clients, NR_CLIENTS); in run_test()
|
/linux-6.12.1/Documentation/admin-guide/nfs/ |
D | pnfs-scsi-server.rst | 9 also hands out layouts to the clients so that they can directly access the 14 file system must sit on a SCSI LUN that is accessible to the clients in 16 exported LUN, striping or concatenation of LUNs on the MDS and clients
|
D | pnfs-block-server.rst | 8 to the clients to directly access the underlying block devices that are 14 to the clients in addition to the MDS. As of now the file system needs to 16 volumes on the MDS and clients is not supported yet.
|
/linux-6.12.1/include/linux/hsi/ |
D | hsi.h | 58 * struct hsi_channel - channel resource used by the hsi clients 210 * @shared: Set when port can be shared by different clients 211 * @claimed: Reference count of clients which claimed the port 219 * @n_head: Notifier chain for signaling port events to the clients. 290 struct device_node *clients); 293 struct device_node *clients) in hsi_add_clients_from_dt() argument 317 * API for HSI clients 349 * When sharing ports, clients should either relay on a single
|
/linux-6.12.1/include/soc/qcom/ |
D | ocmem.h | 3 * The On Chip Memory (OCMEM) allocator allows various clients to allocate 19 /* GMEM clients */ 23 * deal with multiple clients.
|
/linux-6.12.1/drivers/memory/tegra/ |
D | tegra20.c | 378 * ISO clients need to reserve extra bandwidth up-front because in tegra20_mc_icc_aggreate() 410 /* these clients are isochronous by default */ in tegra20_mc_of_icc_xlate_extended() 421 if (mc->soc->clients[i].id == idx) in tegra20_mc_of_icc_xlate_extended() 519 client0 = &mc->soc->clients[i]; in tegra20_mc_collect_stats() 520 client1 = &mc->soc->clients[i + 1]; in tegra20_mc_collect_stats() 533 /* collect more info from active clients */ in tegra20_mc_collect_stats() 539 client0 = &mc->soc->clients[i]; in tegra20_mc_collect_stats() 547 client1 = &mc->soc->clients[i]; in tegra20_mc_collect_stats() 632 seq_printf(s, "%-14s ", mc->soc->clients[i].name); in tegra20_mc_stats_show() 750 client = mc->soc->clients[id].name; in tegra20_mc_handle_irq() [all …]
|
/linux-6.12.1/drivers/misc/mei/ |
D | bus-fixup.c | 51 * We support only clients: 53 * 2. and fixed clients (max_number_of_connections == 0) 55 * @cldev: me clients device 66 * @cldev: me clients device 76 * @cldev: me clients device 297 * @cldev: me clients device 522 * vt_support - enable on bus clients with vtag support 524 * @cldev: me clients device 535 * @cldev: me clients device
|