blob: 03f8e98e3bcc94b781bc66928d7e5e9c667dbe8b [file] [log] [blame]
David Brazdil0f672f62019-12-10 10:32:29 +00001/* SPDX-License-Identifier: GPL-2.0-or-later */
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002#ifndef __LINUX_CMA_H
3#define __LINUX_CMA_H
4
5/*
6 * Contiguous Memory Allocator for DMA mapping framework
7 * Copyright (c) 2010-2011 by Samsung Electronics.
8 * Written by:
9 * Marek Szyprowski <m.szyprowski@samsung.com>
10 * Michal Nazarewicz <mina86@mina86.com>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000011 */
12
13/*
14 * Contiguous Memory Allocator
15 *
16 * The Contiguous Memory Allocator (CMA) makes it possible to
17 * allocate big contiguous chunks of memory after the system has
18 * booted.
19 *
20 * Why is it needed?
21 *
22 * Various devices on embedded systems have no scatter-getter and/or
23 * IO map support and require contiguous blocks of memory to
24 * operate. They include devices such as cameras, hardware video
25 * coders, etc.
26 *
27 * Such devices often require big memory buffers (a full HD frame
28 * is, for instance, more then 2 mega pixels large, i.e. more than 6
29 * MB of memory), which makes mechanisms such as kmalloc() or
30 * alloc_page() ineffective.
31 *
32 * At the same time, a solution where a big memory region is
33 * reserved for a device is suboptimal since often more memory is
34 * reserved then strictly required and, moreover, the memory is
35 * inaccessible to page system even if device drivers don't use it.
36 *
37 * CMA tries to solve this issue by operating on memory regions
38 * where only movable pages can be allocated from. This way, kernel
39 * can use the memory for pagecache and when device driver requests
40 * it, allocated pages can be migrated.
41 *
42 * Driver usage
43 *
44 * CMA should not be used by the device drivers directly. It is
45 * only a helper framework for dma-mapping subsystem.
46 *
47 * For more information, see kernel-docs in kernel/dma/contiguous.c
48 */
49
50#ifdef __KERNEL__
51
52#include <linux/device.h>
David Brazdil0f672f62019-12-10 10:32:29 +000053#include <linux/mm.h>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000054
55struct cma;
56struct page;
57
58#ifdef CONFIG_DMA_CMA
59
60extern struct cma *dma_contiguous_default_area;
61
62static inline struct cma *dev_get_cma_area(struct device *dev)
63{
64 if (dev && dev->cma_area)
65 return dev->cma_area;
66 return dma_contiguous_default_area;
67}
68
69static inline void dev_set_cma_area(struct device *dev, struct cma *cma)
70{
71 if (dev)
72 dev->cma_area = cma;
73}
74
75static inline void dma_contiguous_set_default(struct cma *cma)
76{
77 dma_contiguous_default_area = cma;
78}
79
80void dma_contiguous_reserve(phys_addr_t addr_limit);
81
82int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
83 phys_addr_t limit, struct cma **res_cma,
84 bool fixed);
85
86/**
87 * dma_declare_contiguous() - reserve area for contiguous memory handling
88 * for particular device
89 * @dev: Pointer to device structure.
90 * @size: Size of the reserved memory.
91 * @base: Start address of the reserved memory (optional, 0 for any).
92 * @limit: End address of the reserved memory (optional, 0 for any).
93 *
94 * This function reserves memory for specified device. It should be
95 * called by board specific code when early allocator (memblock or bootmem)
96 * is still activate.
97 */
98
99static inline int dma_declare_contiguous(struct device *dev, phys_addr_t size,
100 phys_addr_t base, phys_addr_t limit)
101{
102 struct cma *cma;
103 int ret;
104 ret = dma_contiguous_reserve_area(size, base, limit, &cma, true);
105 if (ret == 0)
106 dev_set_cma_area(dev, cma);
107
108 return ret;
109}
110
111struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
112 unsigned int order, bool no_warn);
113bool dma_release_from_contiguous(struct device *dev, struct page *pages,
114 int count);
David Brazdil0f672f62019-12-10 10:32:29 +0000115struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
116void dma_free_contiguous(struct device *dev, struct page *page, size_t size);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000117
118#else
119
120static inline struct cma *dev_get_cma_area(struct device *dev)
121{
122 return NULL;
123}
124
125static inline void dev_set_cma_area(struct device *dev, struct cma *cma) { }
126
127static inline void dma_contiguous_set_default(struct cma *cma) { }
128
129static inline void dma_contiguous_reserve(phys_addr_t limit) { }
130
131static inline int dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base,
132 phys_addr_t limit, struct cma **res_cma,
133 bool fixed)
134{
135 return -ENOSYS;
136}
137
138static inline
139int dma_declare_contiguous(struct device *dev, phys_addr_t size,
140 phys_addr_t base, phys_addr_t limit)
141{
142 return -ENOSYS;
143}
144
145static inline
146struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
147 unsigned int order, bool no_warn)
148{
149 return NULL;
150}
151
152static inline
153bool dma_release_from_contiguous(struct device *dev, struct page *pages,
154 int count)
155{
156 return false;
157}
158
David Brazdil0f672f62019-12-10 10:32:29 +0000159/* Use fallback alloc() and free() when CONFIG_DMA_CMA=n */
160static inline struct page *dma_alloc_contiguous(struct device *dev, size_t size,
161 gfp_t gfp)
162{
163 return NULL;
164}
165
166static inline void dma_free_contiguous(struct device *dev, struct page *page,
167 size_t size)
168{
169 __free_pages(page, get_order(size));
170}
171
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000172#endif
173
174#endif
175
176#endif