1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24#define nvkm_vram(p) container_of((p), struct nvkm_vram, memory)
25#include "ram.h"
26
27#include <core/memory.h>
28#include <subdev/mmu.h>
29
30struct nvkm_vram {
31 struct nvkm_memory memory;
32 struct nvkm_ram *ram;
33 u8 page;
34 struct nvkm_mm_node *mn;
35};
36
37static int
38nvkm_vram_map(struct nvkm_memory *memory, u64 offset, struct nvkm_vmm *vmm,
39 struct nvkm_vma *vma, void *argv, u32 argc)
40{
41 struct nvkm_vram *vram = nvkm_vram(memory);
42 struct nvkm_vmm_map map = {
43 .memory = &vram->memory,
44 .offset = offset,
45 .mem = vram->mn,
46 };
47
48 return nvkm_vmm_map(vmm, vma, argv, argc, &map);
49}
50
51static u64
52nvkm_vram_size(struct nvkm_memory *memory)
53{
54 return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT;
55}
56
57static u64
58nvkm_vram_addr(struct nvkm_memory *memory)
59{
60 struct nvkm_vram *vram = nvkm_vram(memory);
61 if (!nvkm_mm_contiguous(vram->mn))
62 return ~0ULL;
63 return (u64)nvkm_mm_addr(vram->mn) << NVKM_RAM_MM_SHIFT;
64}
65
66static u8
67nvkm_vram_page(struct nvkm_memory *memory)
68{
69 return nvkm_vram(memory)->page;
70}
71
72static enum nvkm_memory_target
73nvkm_vram_target(struct nvkm_memory *memory)
74{
75 return NVKM_MEM_TARGET_VRAM;
76}
77
78static void *
79nvkm_vram_dtor(struct nvkm_memory *memory)
80{
81 struct nvkm_vram *vram = nvkm_vram(memory);
82 struct nvkm_mm_node *next = vram->mn;
83 struct nvkm_mm_node *node;
84 mutex_lock(&vram->ram->mutex);
85 while ((node = next)) {
86 next = node->next;
87 nvkm_mm_free(&vram->ram->vram, &node);
88 }
89 mutex_unlock(&vram->ram->mutex);
90 return vram;
91}
92
93static const struct nvkm_memory_func
94nvkm_vram = {
95 .dtor = nvkm_vram_dtor,
96 .target = nvkm_vram_target,
97 .page = nvkm_vram_page,
98 .addr = nvkm_vram_addr,
99 .size = nvkm_vram_size,
100 .map = nvkm_vram_map,
101};
102
103int
104nvkm_ram_get(struct nvkm_device *device, u8 heap, u8 type, u8 rpage, u64 size,
105 bool contig, bool back, struct nvkm_memory **pmemory)
106{
107 struct nvkm_ram *ram;
108 struct nvkm_mm *mm;
109 struct nvkm_mm_node **node, *r;
110 struct nvkm_vram *vram;
111 u8 page = max(rpage, (u8)NVKM_RAM_MM_SHIFT);
112 u32 align = (1 << page) >> NVKM_RAM_MM_SHIFT;
113 u32 max = ALIGN(size, 1 << page) >> NVKM_RAM_MM_SHIFT;
114 u32 min = contig ? max : align;
115 int ret;
116
117 if (!device->fb || !(ram = device->fb->ram))
118 return -ENODEV;
119 ram = device->fb->ram;
120 mm = &ram->vram;
121
122 if (!(vram = kzalloc(sizeof(*vram), GFP_KERNEL)))
123 return -ENOMEM;
124 nvkm_memory_ctor(&nvkm_vram, &vram->memory);
125 vram->ram = ram;
126 vram->page = page;
127 *pmemory = &vram->memory;
128
129 mutex_lock(&ram->mutex);
130 node = &vram->mn;
131 do {
132 if (back)
133 ret = nvkm_mm_tail(mm, heap, type, max, min, align, &r);
134 else
135 ret = nvkm_mm_head(mm, heap, type, max, min, align, &r);
136 if (ret) {
137 mutex_unlock(&ram->mutex);
138 nvkm_memory_unref(pmemory);
139 return ret;
140 }
141
142 *node = r;
143 node = &r->next;
144 max -= r->length;
145 } while (max);
146 mutex_unlock(&ram->mutex);
147 return 0;
148}
149
150int
151nvkm_ram_init(struct nvkm_ram *ram)
152{
153 if (ram->func->init)
154 return ram->func->init(ram);
155 return 0;
156}
157
158void
159nvkm_ram_del(struct nvkm_ram **pram)
160{
161 struct nvkm_ram *ram = *pram;
162 if (ram && !WARN_ON(!ram->func)) {
163 if (ram->func->dtor)
164 *pram = ram->func->dtor(ram);
165 nvkm_mm_fini(&ram->vram);
166 mutex_destroy(&ram->mutex);
167 kfree(*pram);
168 *pram = NULL;
169 }
170}
171
172int
173nvkm_ram_ctor(const struct nvkm_ram_func *func, struct nvkm_fb *fb,
174 enum nvkm_ram_type type, u64 size, struct nvkm_ram *ram)
175{
176 static const char *name[] = {
177 [NVKM_RAM_TYPE_UNKNOWN] = "of unknown memory type",
178 [NVKM_RAM_TYPE_STOLEN ] = "stolen system memory",
179 [NVKM_RAM_TYPE_SGRAM ] = "SGRAM",
180 [NVKM_RAM_TYPE_SDRAM ] = "SDRAM",
181 [NVKM_RAM_TYPE_DDR1 ] = "DDR1",
182 [NVKM_RAM_TYPE_DDR2 ] = "DDR2",
183 [NVKM_RAM_TYPE_DDR3 ] = "DDR3",
184 [NVKM_RAM_TYPE_GDDR2 ] = "GDDR2",
185 [NVKM_RAM_TYPE_GDDR3 ] = "GDDR3",
186 [NVKM_RAM_TYPE_GDDR4 ] = "GDDR4",
187 [NVKM_RAM_TYPE_GDDR5 ] = "GDDR5",
188 [NVKM_RAM_TYPE_GDDR5X ] = "GDDR5X",
189 [NVKM_RAM_TYPE_GDDR6 ] = "GDDR6",
190 [NVKM_RAM_TYPE_HBM2 ] = "HBM2",
191 };
192 struct nvkm_subdev *subdev = &fb->subdev;
193 int ret;
194
195 nvkm_info(subdev, "%d MiB %s\n", (int)(size >> 20), name[type]);
196 ram->func = func;
197 ram->fb = fb;
198 ram->type = type;
199 ram->size = size;
200 mutex_init(&ram->mutex);
201
202 if (!nvkm_mm_initialised(&ram->vram)) {
203 ret = nvkm_mm_init(&ram->vram, NVKM_RAM_MM_NORMAL, 0,
204 size >> NVKM_RAM_MM_SHIFT, 1);
205 if (ret)
206 return ret;
207 }
208
209 return 0;
210}
211
212int
213nvkm_ram_new_(const struct nvkm_ram_func *func, struct nvkm_fb *fb,
214 enum nvkm_ram_type type, u64 size, struct nvkm_ram **pram)
215{
216 if (!(*pram = kzalloc(sizeof(**pram), GFP_KERNEL)))
217 return -ENOMEM;
218 return nvkm_ram_ctor(func, fb, type, size, *pram);
219}
220