Marc-André Lureau | d52c454 | 2019-05-24 15:09:42 +0200 | [diff] [blame] | 1 | /* |
| 2 | * Virtio vhost-user GPU Device |
| 3 | * |
| 4 | * DRM helpers |
| 5 | * |
| 6 | * This work is licensed under the terms of the GNU GPL, version 2 or later. |
| 7 | * See the COPYING file in the top-level directory. |
| 8 | */ |
| 9 | |
| 10 | #include "vugbm.h" |
| 11 | |
| 12 | static bool |
| 13 | mem_alloc_bo(struct vugbm_buffer *buf) |
| 14 | { |
| 15 | buf->mmap = g_malloc(buf->width * buf->height * 4); |
| 16 | buf->stride = buf->width * 4; |
| 17 | return true; |
| 18 | } |
| 19 | |
| 20 | static void |
| 21 | mem_free_bo(struct vugbm_buffer *buf) |
| 22 | { |
| 23 | g_free(buf->mmap); |
| 24 | } |
| 25 | |
| 26 | static bool |
| 27 | mem_map_bo(struct vugbm_buffer *buf) |
| 28 | { |
| 29 | return buf->mmap != NULL; |
| 30 | } |
| 31 | |
| 32 | static void |
| 33 | mem_unmap_bo(struct vugbm_buffer *buf) |
| 34 | { |
| 35 | } |
| 36 | |
| 37 | static void |
| 38 | mem_device_destroy(struct vugbm_device *dev) |
| 39 | { |
| 40 | } |
| 41 | |
| 42 | #ifdef CONFIG_MEMFD |
| 43 | struct udmabuf_create { |
| 44 | uint32_t memfd; |
| 45 | uint32_t flags; |
| 46 | uint64_t offset; |
| 47 | uint64_t size; |
| 48 | }; |
| 49 | |
| 50 | #define UDMABUF_CREATE _IOW('u', 0x42, struct udmabuf_create) |
| 51 | |
| 52 | static size_t |
| 53 | udmabuf_get_size(struct vugbm_buffer *buf) |
| 54 | { |
Wei Yang | 038adc2 | 2019-10-13 10:11:45 +0800 | [diff] [blame] | 55 | return ROUND_UP(buf->width * buf->height * 4, qemu_real_host_page_size); |
Marc-André Lureau | d52c454 | 2019-05-24 15:09:42 +0200 | [diff] [blame] | 56 | } |
| 57 | |
| 58 | static bool |
| 59 | udmabuf_alloc_bo(struct vugbm_buffer *buf) |
| 60 | { |
| 61 | int ret; |
| 62 | |
| 63 | buf->memfd = memfd_create("udmabuf-bo", MFD_ALLOW_SEALING); |
| 64 | if (buf->memfd < 0) { |
| 65 | return false; |
| 66 | } |
| 67 | |
| 68 | ret = ftruncate(buf->memfd, udmabuf_get_size(buf)); |
| 69 | if (ret < 0) { |
| 70 | close(buf->memfd); |
| 71 | return false; |
| 72 | } |
| 73 | |
| 74 | ret = fcntl(buf->memfd, F_ADD_SEALS, F_SEAL_SHRINK); |
| 75 | if (ret < 0) { |
| 76 | close(buf->memfd); |
| 77 | return false; |
| 78 | } |
| 79 | |
| 80 | buf->stride = buf->width * 4; |
| 81 | |
| 82 | return true; |
| 83 | } |
| 84 | |
| 85 | static void |
| 86 | udmabuf_free_bo(struct vugbm_buffer *buf) |
| 87 | { |
| 88 | close(buf->memfd); |
| 89 | } |
| 90 | |
| 91 | static bool |
| 92 | udmabuf_map_bo(struct vugbm_buffer *buf) |
| 93 | { |
| 94 | buf->mmap = mmap(NULL, udmabuf_get_size(buf), |
| 95 | PROT_READ | PROT_WRITE, MAP_SHARED, buf->memfd, 0); |
| 96 | if (buf->mmap == MAP_FAILED) { |
| 97 | return false; |
| 98 | } |
| 99 | |
| 100 | return true; |
| 101 | } |
| 102 | |
| 103 | static bool |
| 104 | udmabuf_get_fd(struct vugbm_buffer *buf, int *fd) |
| 105 | { |
| 106 | struct udmabuf_create create = { |
| 107 | .memfd = buf->memfd, |
| 108 | .offset = 0, |
| 109 | .size = udmabuf_get_size(buf), |
| 110 | }; |
| 111 | |
| 112 | *fd = ioctl(buf->dev->fd, UDMABUF_CREATE, &create); |
| 113 | |
| 114 | return *fd >= 0; |
| 115 | } |
| 116 | |
| 117 | static void |
| 118 | udmabuf_unmap_bo(struct vugbm_buffer *buf) |
| 119 | { |
| 120 | munmap(buf->mmap, udmabuf_get_size(buf)); |
| 121 | } |
| 122 | |
| 123 | static void |
| 124 | udmabuf_device_destroy(struct vugbm_device *dev) |
| 125 | { |
| 126 | close(dev->fd); |
| 127 | } |
| 128 | #endif |
| 129 | |
| 130 | #ifdef CONFIG_GBM |
| 131 | static bool |
| 132 | alloc_bo(struct vugbm_buffer *buf) |
| 133 | { |
| 134 | struct gbm_device *dev = buf->dev->dev; |
| 135 | |
| 136 | assert(!buf->bo); |
| 137 | |
| 138 | buf->bo = gbm_bo_create(dev, buf->width, buf->height, |
| 139 | buf->format, |
| 140 | GBM_BO_USE_RENDERING | GBM_BO_USE_LINEAR); |
| 141 | |
| 142 | if (buf->bo) { |
| 143 | buf->stride = gbm_bo_get_stride(buf->bo); |
| 144 | return true; |
| 145 | } |
| 146 | |
| 147 | return false; |
| 148 | } |
| 149 | |
| 150 | static void |
| 151 | free_bo(struct vugbm_buffer *buf) |
| 152 | { |
| 153 | gbm_bo_destroy(buf->bo); |
| 154 | } |
| 155 | |
| 156 | static bool |
| 157 | map_bo(struct vugbm_buffer *buf) |
| 158 | { |
| 159 | uint32_t stride; |
| 160 | |
| 161 | buf->mmap = gbm_bo_map(buf->bo, 0, 0, buf->width, buf->height, |
| 162 | GBM_BO_TRANSFER_READ_WRITE, &stride, |
| 163 | &buf->mmap_data); |
| 164 | |
| 165 | assert(stride == buf->stride); |
| 166 | |
| 167 | return buf->mmap != NULL; |
| 168 | } |
| 169 | |
| 170 | static void |
| 171 | unmap_bo(struct vugbm_buffer *buf) |
| 172 | { |
| 173 | gbm_bo_unmap(buf->bo, buf->mmap_data); |
| 174 | } |
| 175 | |
| 176 | static bool |
| 177 | get_fd(struct vugbm_buffer *buf, int *fd) |
| 178 | { |
| 179 | *fd = gbm_bo_get_fd(buf->bo); |
| 180 | |
| 181 | return *fd >= 0; |
| 182 | } |
| 183 | |
| 184 | static void |
| 185 | device_destroy(struct vugbm_device *dev) |
| 186 | { |
| 187 | gbm_device_destroy(dev->dev); |
| 188 | } |
| 189 | #endif |
| 190 | |
| 191 | void |
| 192 | vugbm_device_destroy(struct vugbm_device *dev) |
| 193 | { |
| 194 | if (!dev->inited) { |
| 195 | return; |
| 196 | } |
| 197 | |
| 198 | dev->device_destroy(dev); |
| 199 | } |
| 200 | |
| 201 | bool |
| 202 | vugbm_device_init(struct vugbm_device *dev, int fd) |
| 203 | { |
| 204 | dev->fd = fd; |
| 205 | |
| 206 | #ifdef CONFIG_GBM |
| 207 | dev->dev = gbm_create_device(fd); |
| 208 | #endif |
| 209 | |
| 210 | if (0) { |
| 211 | /* nothing */ |
| 212 | } |
| 213 | #ifdef CONFIG_GBM |
| 214 | else if (dev->dev != NULL) { |
| 215 | dev->alloc_bo = alloc_bo; |
| 216 | dev->free_bo = free_bo; |
| 217 | dev->get_fd = get_fd; |
| 218 | dev->map_bo = map_bo; |
| 219 | dev->unmap_bo = unmap_bo; |
| 220 | dev->device_destroy = device_destroy; |
| 221 | } |
| 222 | #endif |
| 223 | #ifdef CONFIG_MEMFD |
| 224 | else if (g_file_test("/dev/udmabuf", G_FILE_TEST_EXISTS)) { |
| 225 | dev->fd = open("/dev/udmabuf", O_RDWR); |
| 226 | if (dev->fd < 0) { |
| 227 | return false; |
| 228 | } |
| 229 | g_debug("Using experimental udmabuf backend"); |
| 230 | dev->alloc_bo = udmabuf_alloc_bo; |
| 231 | dev->free_bo = udmabuf_free_bo; |
| 232 | dev->get_fd = udmabuf_get_fd; |
| 233 | dev->map_bo = udmabuf_map_bo; |
| 234 | dev->unmap_bo = udmabuf_unmap_bo; |
| 235 | dev->device_destroy = udmabuf_device_destroy; |
| 236 | } |
| 237 | #endif |
| 238 | else { |
| 239 | g_debug("Using mem fallback"); |
| 240 | dev->alloc_bo = mem_alloc_bo; |
| 241 | dev->free_bo = mem_free_bo; |
| 242 | dev->map_bo = mem_map_bo; |
| 243 | dev->unmap_bo = mem_unmap_bo; |
| 244 | dev->device_destroy = mem_device_destroy; |
| 245 | return false; |
| 246 | } |
| 247 | |
| 248 | dev->inited = true; |
| 249 | return true; |
| 250 | } |
| 251 | |
| 252 | static bool |
| 253 | vugbm_buffer_map(struct vugbm_buffer *buf) |
| 254 | { |
| 255 | struct vugbm_device *dev = buf->dev; |
| 256 | |
| 257 | return dev->map_bo(buf); |
| 258 | } |
| 259 | |
| 260 | static void |
| 261 | vugbm_buffer_unmap(struct vugbm_buffer *buf) |
| 262 | { |
| 263 | struct vugbm_device *dev = buf->dev; |
| 264 | |
| 265 | dev->unmap_bo(buf); |
| 266 | } |
| 267 | |
| 268 | bool |
| 269 | vugbm_buffer_can_get_dmabuf_fd(struct vugbm_buffer *buffer) |
| 270 | { |
| 271 | if (!buffer->dev->get_fd) { |
| 272 | return false; |
| 273 | } |
| 274 | |
| 275 | return true; |
| 276 | } |
| 277 | |
| 278 | bool |
| 279 | vugbm_buffer_get_dmabuf_fd(struct vugbm_buffer *buffer, int *fd) |
| 280 | { |
| 281 | if (!vugbm_buffer_can_get_dmabuf_fd(buffer) || |
| 282 | !buffer->dev->get_fd(buffer, fd)) { |
| 283 | g_warning("Failed to get dmabuf"); |
| 284 | return false; |
| 285 | } |
| 286 | |
| 287 | if (*fd < 0) { |
| 288 | g_warning("error: dmabuf_fd < 0"); |
| 289 | return false; |
| 290 | } |
| 291 | |
| 292 | return true; |
| 293 | } |
| 294 | |
| 295 | bool |
| 296 | vugbm_buffer_create(struct vugbm_buffer *buffer, struct vugbm_device *dev, |
| 297 | uint32_t width, uint32_t height) |
| 298 | { |
| 299 | buffer->dev = dev; |
| 300 | buffer->width = width; |
| 301 | buffer->height = height; |
| 302 | buffer->format = GBM_FORMAT_XRGB8888; |
| 303 | buffer->stride = 0; /* modified during alloc */ |
| 304 | if (!dev->alloc_bo(buffer)) { |
| 305 | g_warning("alloc_bo failed"); |
| 306 | return false; |
| 307 | } |
| 308 | |
| 309 | if (!vugbm_buffer_map(buffer)) { |
| 310 | g_warning("map_bo failed"); |
| 311 | goto err; |
| 312 | } |
| 313 | |
| 314 | return true; |
| 315 | |
| 316 | err: |
| 317 | dev->free_bo(buffer); |
| 318 | return false; |
| 319 | } |
| 320 | |
| 321 | void |
| 322 | vugbm_buffer_destroy(struct vugbm_buffer *buffer) |
| 323 | { |
| 324 | struct vugbm_device *dev = buffer->dev; |
| 325 | |
| 326 | vugbm_buffer_unmap(buffer); |
| 327 | dev->free_bo(buffer); |
| 328 | } |