1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * From setup-res.c, by:
4  *	Dave Rusling (david.rusling@reo.mts.dec.com)
5  *	David Mosberger (davidm@cs.arizona.edu)
6  *	David Miller (davem@redhat.com)
7  *	Ivan Kokshaysky (ink@jurassic.park.msu.ru)
8  */
9 #include <linux/module.h>
10 #include <linux/kernel.h>
11 #include <linux/pci.h>
12 #include <linux/errno.h>
13 #include <linux/ioport.h>
14 #include <linux/of.h>
15 #include <linux/of_platform.h>
16 #include <linux/proc_fs.h>
17 #include <linux/slab.h>
18 
19 #include "pci.h"
20 
pci_add_resource_offset(struct list_head * resources,struct resource * res,resource_size_t offset)21 void pci_add_resource_offset(struct list_head *resources, struct resource *res,
22 			     resource_size_t offset)
23 {
24 	struct resource_entry *entry;
25 
26 	entry = resource_list_create_entry(res, 0);
27 	if (!entry) {
28 		pr_err("PCI: can't add host bridge window %pR\n", res);
29 		return;
30 	}
31 
32 	entry->offset = offset;
33 	resource_list_add_tail(entry, resources);
34 }
35 EXPORT_SYMBOL(pci_add_resource_offset);
36 
pci_add_resource(struct list_head * resources,struct resource * res)37 void pci_add_resource(struct list_head *resources, struct resource *res)
38 {
39 	pci_add_resource_offset(resources, res, 0);
40 }
41 EXPORT_SYMBOL(pci_add_resource);
42 
pci_free_resource_list(struct list_head * resources)43 void pci_free_resource_list(struct list_head *resources)
44 {
45 	resource_list_free(resources);
46 }
47 EXPORT_SYMBOL(pci_free_resource_list);
48 
pci_bus_add_resource(struct pci_bus * bus,struct resource * res,unsigned int flags)49 void pci_bus_add_resource(struct pci_bus *bus, struct resource *res,
50 			  unsigned int flags)
51 {
52 	struct pci_bus_resource *bus_res;
53 
54 	bus_res = kzalloc(sizeof(struct pci_bus_resource), GFP_KERNEL);
55 	if (!bus_res) {
56 		dev_err(&bus->dev, "can't add %pR resource\n", res);
57 		return;
58 	}
59 
60 	bus_res->res = res;
61 	bus_res->flags = flags;
62 	list_add_tail(&bus_res->list, &bus->resources);
63 }
64 
pci_bus_resource_n(const struct pci_bus * bus,int n)65 struct resource *pci_bus_resource_n(const struct pci_bus *bus, int n)
66 {
67 	struct pci_bus_resource *bus_res;
68 
69 	if (n < PCI_BRIDGE_RESOURCE_NUM)
70 		return bus->resource[n];
71 
72 	n -= PCI_BRIDGE_RESOURCE_NUM;
73 	list_for_each_entry(bus_res, &bus->resources, list) {
74 		if (n-- == 0)
75 			return bus_res->res;
76 	}
77 	return NULL;
78 }
79 EXPORT_SYMBOL_GPL(pci_bus_resource_n);
80 
pci_bus_remove_resource(struct pci_bus * bus,struct resource * res)81 void pci_bus_remove_resource(struct pci_bus *bus, struct resource *res)
82 {
83 	struct pci_bus_resource *bus_res, *tmp;
84 	int i;
85 
86 	for (i = 0; i < PCI_BRIDGE_RESOURCE_NUM; i++) {
87 		if (bus->resource[i] == res) {
88 			bus->resource[i] = NULL;
89 			return;
90 		}
91 	}
92 
93 	list_for_each_entry_safe(bus_res, tmp, &bus->resources, list) {
94 		if (bus_res->res == res) {
95 			list_del(&bus_res->list);
96 			kfree(bus_res);
97 			return;
98 		}
99 	}
100 }
101 
pci_bus_remove_resources(struct pci_bus * bus)102 void pci_bus_remove_resources(struct pci_bus *bus)
103 {
104 	int i;
105 	struct pci_bus_resource *bus_res, *tmp;
106 
107 	for (i = 0; i < PCI_BRIDGE_RESOURCE_NUM; i++)
108 		bus->resource[i] = NULL;
109 
110 	list_for_each_entry_safe(bus_res, tmp, &bus->resources, list) {
111 		list_del(&bus_res->list);
112 		kfree(bus_res);
113 	}
114 }
115 
devm_request_pci_bus_resources(struct device * dev,struct list_head * resources)116 int devm_request_pci_bus_resources(struct device *dev,
117 				   struct list_head *resources)
118 {
119 	struct resource_entry *win;
120 	struct resource *parent, *res;
121 	int err;
122 
123 	resource_list_for_each_entry(win, resources) {
124 		res = win->res;
125 		switch (resource_type(res)) {
126 		case IORESOURCE_IO:
127 			parent = &ioport_resource;
128 			break;
129 		case IORESOURCE_MEM:
130 			parent = &iomem_resource;
131 			break;
132 		default:
133 			continue;
134 		}
135 
136 		err = devm_request_resource(dev, parent, res);
137 		if (err)
138 			return err;
139 	}
140 
141 	return 0;
142 }
143 EXPORT_SYMBOL_GPL(devm_request_pci_bus_resources);
144 
145 static struct pci_bus_region pci_32_bit = {0, 0xffffffffULL};
146 #ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
147 static struct pci_bus_region pci_64_bit = {0,
148 				(pci_bus_addr_t) 0xffffffffffffffffULL};
149 static struct pci_bus_region pci_high = {(pci_bus_addr_t) 0x100000000ULL,
150 				(pci_bus_addr_t) 0xffffffffffffffffULL};
151 #endif
152 
153 /*
154  * @res contains CPU addresses.  Clip it so the corresponding bus addresses
155  * on @bus are entirely within @region.  This is used to control the bus
156  * addresses of resources we allocate, e.g., we may need a resource that
157  * can be mapped by a 32-bit BAR.
158  */
pci_clip_resource_to_region(struct pci_bus * bus,struct resource * res,struct pci_bus_region * region)159 static void pci_clip_resource_to_region(struct pci_bus *bus,
160 					struct resource *res,
161 					struct pci_bus_region *region)
162 {
163 	struct pci_bus_region r;
164 
165 	pcibios_resource_to_bus(bus, &r, res);
166 	if (r.start < region->start)
167 		r.start = region->start;
168 	if (r.end > region->end)
169 		r.end = region->end;
170 
171 	if (r.end < r.start)
172 		res->end = res->start - 1;
173 	else
174 		pcibios_bus_to_resource(bus, res, &r);
175 }
176 
pci_bus_alloc_from_region(struct pci_bus * bus,struct resource * res,resource_size_t size,resource_size_t align,resource_size_t min,unsigned long type_mask,resource_alignf alignf,void * alignf_data,struct pci_bus_region * region)177 static int pci_bus_alloc_from_region(struct pci_bus *bus, struct resource *res,
178 		resource_size_t size, resource_size_t align,
179 		resource_size_t min, unsigned long type_mask,
180 		resource_alignf alignf,
181 		void *alignf_data,
182 		struct pci_bus_region *region)
183 {
184 	struct resource *r, avail;
185 	resource_size_t max;
186 	int ret;
187 
188 	type_mask |= IORESOURCE_TYPE_BITS;
189 
190 	pci_bus_for_each_resource(bus, r) {
191 		resource_size_t min_used = min;
192 
193 		if (!r)
194 			continue;
195 
196 		/* type_mask must match */
197 		if ((res->flags ^ r->flags) & type_mask)
198 			continue;
199 
200 		/* We cannot allocate a non-prefetching resource
201 		   from a pre-fetching area */
202 		if ((r->flags & IORESOURCE_PREFETCH) &&
203 		    !(res->flags & IORESOURCE_PREFETCH))
204 			continue;
205 
206 		avail = *r;
207 		pci_clip_resource_to_region(bus, &avail, region);
208 
209 		/*
210 		 * "min" is typically PCIBIOS_MIN_IO or PCIBIOS_MIN_MEM to
211 		 * protect badly documented motherboard resources, but if
212 		 * this is an already-configured bridge window, its start
213 		 * overrides "min".
214 		 */
215 		if (avail.start)
216 			min_used = avail.start;
217 
218 		max = avail.end;
219 
220 		/* Don't bother if available space isn't large enough */
221 		if (size > max - min_used + 1)
222 			continue;
223 
224 		/* Ok, try it out.. */
225 		ret = allocate_resource(r, res, size, min_used, max,
226 					align, alignf, alignf_data);
227 		if (ret == 0)
228 			return 0;
229 	}
230 	return -ENOMEM;
231 }
232 
233 /**
234  * pci_bus_alloc_resource - allocate a resource from a parent bus
235  * @bus: PCI bus
236  * @res: resource to allocate
237  * @size: size of resource to allocate
238  * @align: alignment of resource to allocate
239  * @min: minimum /proc/iomem address to allocate
240  * @type_mask: IORESOURCE_* type flags
241  * @alignf: resource alignment function
242  * @alignf_data: data argument for resource alignment function
243  *
244  * Given the PCI bus a device resides on, the size, minimum address,
245  * alignment and type, try to find an acceptable resource allocation
246  * for a specific device resource.
247  */
pci_bus_alloc_resource(struct pci_bus * bus,struct resource * res,resource_size_t size,resource_size_t align,resource_size_t min,unsigned long type_mask,resource_alignf alignf,void * alignf_data)248 int pci_bus_alloc_resource(struct pci_bus *bus, struct resource *res,
249 		resource_size_t size, resource_size_t align,
250 		resource_size_t min, unsigned long type_mask,
251 		resource_alignf alignf,
252 		void *alignf_data)
253 {
254 #ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
255 	int rc;
256 
257 	if (res->flags & IORESOURCE_MEM_64) {
258 		rc = pci_bus_alloc_from_region(bus, res, size, align, min,
259 					       type_mask, alignf, alignf_data,
260 					       &pci_high);
261 		if (rc == 0)
262 			return 0;
263 
264 		return pci_bus_alloc_from_region(bus, res, size, align, min,
265 						 type_mask, alignf, alignf_data,
266 						 &pci_64_bit);
267 	}
268 #endif
269 
270 	return pci_bus_alloc_from_region(bus, res, size, align, min,
271 					 type_mask, alignf, alignf_data,
272 					 &pci_32_bit);
273 }
274 EXPORT_SYMBOL(pci_bus_alloc_resource);
275 
276 /*
277  * The @idx resource of @dev should be a PCI-PCI bridge window.  If this
278  * resource fits inside a window of an upstream bridge, do nothing.  If it
279  * overlaps an upstream window but extends outside it, clip the resource so
280  * it fits completely inside.
281  */
pci_bus_clip_resource(struct pci_dev * dev,int idx)282 bool pci_bus_clip_resource(struct pci_dev *dev, int idx)
283 {
284 	struct pci_bus *bus = dev->bus;
285 	struct resource *res = &dev->resource[idx];
286 	struct resource orig_res = *res;
287 	struct resource *r;
288 
289 	pci_bus_for_each_resource(bus, r) {
290 		resource_size_t start, end;
291 
292 		if (!r)
293 			continue;
294 
295 		if (resource_type(res) != resource_type(r))
296 			continue;
297 
298 		start = max(r->start, res->start);
299 		end = min(r->end, res->end);
300 
301 		if (start > end)
302 			continue;	/* no overlap */
303 
304 		if (res->start == start && res->end == end)
305 			return false;	/* no change */
306 
307 		res->start = start;
308 		res->end = end;
309 		res->flags &= ~IORESOURCE_UNSET;
310 		orig_res.flags &= ~IORESOURCE_UNSET;
311 		pci_info(dev, "%pR clipped to %pR\n", &orig_res, res);
312 
313 		return true;
314 	}
315 
316 	return false;
317 }
318 
pcibios_resource_survey_bus(struct pci_bus * bus)319 void __weak pcibios_resource_survey_bus(struct pci_bus *bus) { }
320 
pcibios_bus_add_device(struct pci_dev * pdev)321 void __weak pcibios_bus_add_device(struct pci_dev *pdev) { }
322 
323 /**
324  * pci_bus_add_device - start driver for a single device
325  * @dev: device to add
326  *
327  * This adds add sysfs entries and start device drivers
328  */
pci_bus_add_device(struct pci_dev * dev)329 void pci_bus_add_device(struct pci_dev *dev)
330 {
331 	struct device_node *dn = dev->dev.of_node;
332 	int retval;
333 
334 	/*
335 	 * Can not put in pci_device_add yet because resources
336 	 * are not assigned yet for some devices.
337 	 */
338 	pcibios_bus_add_device(dev);
339 	pci_fixup_device(pci_fixup_final, dev);
340 	if (pci_is_bridge(dev))
341 		of_pci_make_dev_node(dev);
342 	pci_create_sysfs_dev_files(dev);
343 	pci_proc_attach_device(dev);
344 	pci_bridge_d3_update(dev);
345 
346 	dev->match_driver = !dn || of_device_is_available(dn);
347 	retval = device_attach(&dev->dev);
348 	if (retval < 0 && retval != -EPROBE_DEFER)
349 		pci_warn(dev, "device attach failed (%d)\n", retval);
350 
351 	pci_dev_assign_added(dev, true);
352 
353 	if (dev_of_node(&dev->dev) && pci_is_bridge(dev)) {
354 		retval = of_platform_populate(dev_of_node(&dev->dev), NULL, NULL,
355 					      &dev->dev);
356 		if (retval)
357 			pci_err(dev, "failed to populate child OF nodes (%d)\n",
358 				retval);
359 	}
360 }
361 EXPORT_SYMBOL_GPL(pci_bus_add_device);
362 
363 /**
364  * pci_bus_add_devices - start driver for PCI devices
365  * @bus: bus to check for new devices
366  *
367  * Start driver for PCI devices and add some sysfs entries.
368  */
pci_bus_add_devices(const struct pci_bus * bus)369 void pci_bus_add_devices(const struct pci_bus *bus)
370 {
371 	struct pci_dev *dev;
372 	struct pci_bus *child;
373 
374 	list_for_each_entry(dev, &bus->devices, bus_list) {
375 		/* Skip already-added devices */
376 		if (pci_dev_is_added(dev))
377 			continue;
378 		pci_bus_add_device(dev);
379 	}
380 
381 	list_for_each_entry(dev, &bus->devices, bus_list) {
382 		/* Skip if device attach failed */
383 		if (!pci_dev_is_added(dev))
384 			continue;
385 		child = dev->subordinate;
386 		if (child)
387 			pci_bus_add_devices(child);
388 	}
389 }
390 EXPORT_SYMBOL(pci_bus_add_devices);
391 
__pci_walk_bus(struct pci_bus * top,int (* cb)(struct pci_dev *,void *),void * userdata,bool locked)392 static void __pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *),
393 			   void *userdata, bool locked)
394 {
395 	struct pci_dev *dev;
396 	struct pci_bus *bus;
397 	struct list_head *next;
398 	int retval;
399 
400 	bus = top;
401 	if (!locked)
402 		down_read(&pci_bus_sem);
403 	next = top->devices.next;
404 	for (;;) {
405 		if (next == &bus->devices) {
406 			/* end of this bus, go up or finish */
407 			if (bus == top)
408 				break;
409 			next = bus->self->bus_list.next;
410 			bus = bus->self->bus;
411 			continue;
412 		}
413 		dev = list_entry(next, struct pci_dev, bus_list);
414 		if (dev->subordinate) {
415 			/* this is a pci-pci bridge, do its devices next */
416 			next = dev->subordinate->devices.next;
417 			bus = dev->subordinate;
418 		} else
419 			next = dev->bus_list.next;
420 
421 		retval = cb(dev, userdata);
422 		if (retval)
423 			break;
424 	}
425 	if (!locked)
426 		up_read(&pci_bus_sem);
427 }
428 
429 /**
430  *  pci_walk_bus - walk devices on/under bus, calling callback.
431  *  @top: bus whose devices should be walked
432  *  @cb: callback to be called for each device found
433  *  @userdata: arbitrary pointer to be passed to callback
434  *
435  *  Walk the given bus, including any bridged devices
436  *  on buses under this bus.  Call the provided callback
437  *  on each device found.
438  *
439  *  We check the return of @cb each time. If it returns anything
440  *  other than 0, we break out.
441  */
pci_walk_bus(struct pci_bus * top,int (* cb)(struct pci_dev *,void *),void * userdata)442 void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *), void *userdata)
443 {
444 	__pci_walk_bus(top, cb, userdata, false);
445 }
446 EXPORT_SYMBOL_GPL(pci_walk_bus);
447 
pci_walk_bus_locked(struct pci_bus * top,int (* cb)(struct pci_dev *,void *),void * userdata)448 void pci_walk_bus_locked(struct pci_bus *top, int (*cb)(struct pci_dev *, void *), void *userdata)
449 {
450 	lockdep_assert_held(&pci_bus_sem);
451 
452 	__pci_walk_bus(top, cb, userdata, true);
453 }
454 EXPORT_SYMBOL_GPL(pci_walk_bus_locked);
455 
pci_bus_get(struct pci_bus * bus)456 struct pci_bus *pci_bus_get(struct pci_bus *bus)
457 {
458 	if (bus)
459 		get_device(&bus->dev);
460 	return bus;
461 }
462 
pci_bus_put(struct pci_bus * bus)463 void pci_bus_put(struct pci_bus *bus)
464 {
465 	if (bus)
466 		put_device(&bus->dev);
467 }
468