blob: bbdc80a7f45426f6b72a5e3f22df47a6e1f76295 [file] [log] [blame]
/*
* Copyright © 2008-2011 Kristian Høgsberg
* Copyright © 2011 Intel Corporation
* Copyright © 2017, 2018 Collabora, Ltd.
* Copyright © 2017, 2018 General Electric Company
*
* Permission is hereby granted, free of charge, to any person obtaining
* a copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sublicense, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice (including the
* next paragraph) shall be included in all copies or substantial
* portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
#include "config.h"
#include <errno.h>
#include <stdint.h>
#include <stdlib.h>
#include <ctype.h>
#include <string.h>
#include <fcntl.h>
#include <unistd.h>
#include <linux/input.h>
#include <linux/vt.h>
#include <assert.h>
#include <sys/mman.h>
#include <dlfcn.h>
#include <time.h>
#include <xf86drm.h>
#include <xf86drmMode.h>
#include <drm_fourcc.h>
#include <gbm.h>
#include <libudev.h>
#include "compositor/weston.h"
#include "compositor.h"
#include "compositor-drm.h"
#include "shared/helpers.h"
#include "shared/timespec-util.h"
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_OPENGL)
#include "gl-renderer.h"
#endif
#if defined(ENABLE_IMXG2D)
#include "g2d-renderer.h"
#endif
#endif
#include "weston-egl-ext.h"
#include "pixman-renderer.h"
#include "pixel-formats.h"
#include "libbacklight.h"
#include "libinput-seat.h"
#include "launcher-util.h"
#include "vaapi-recorder.h"
#include "presentation-time-server-protocol.h"
#include "linux-dmabuf.h"
#include "linux-dmabuf-unstable-v1-server-protocol.h"
#include "hdr10-metadata-unstable-v1-server-protocol.h"
#ifndef DRM_CLIENT_CAP_ASPECT_RATIO
#define DRM_CLIENT_CAP_ASPECT_RATIO 4
#endif
#ifndef GBM_BO_USE_CURSOR
#define GBM_BO_USE_CURSOR GBM_BO_USE_CURSOR_64X64
#endif
#define MAX_CLONED_CONNECTORS 4
/**
* aspect ratio info taken from the drmModeModeInfo flag bits 19-22,
* which should be used to fill the aspect ratio field in weston_mode.
*/
#define DRM_MODE_FLAG_PIC_AR_BITS_POS 19
#ifndef DRM_MODE_FLAG_PIC_AR_MASK
#define DRM_MODE_FLAG_PIC_AR_MASK (0xF << DRM_MODE_FLAG_PIC_AR_BITS_POS)
#endif
#define ALIGNTO(a, b) ((a + (b-1)) & (~(b-1)))
/**
* Represents the values of an enum-type KMS property
*/
struct drm_property_enum_info {
const char *name; /**< name as string (static, not freed) */
bool valid; /**< true if value is supported; ignore if false */
uint64_t value; /**< raw value */
};
/**
* Holds information on a DRM property, including its ID and the enum
* values it holds.
*
* DRM properties are allocated dynamically, and maintained as DRM objects
* within the normal object ID space; they thus do not have a stable ID
* to refer to. This includes enum values, which must be referred to by
* integer values, but these are not stable.
*
* drm_property_info allows a cache to be maintained where Weston can use
* enum values internally to refer to properties, with the mapping to DRM
* ID values being maintained internally.
*/
struct drm_property_info {
const char *name; /**< name as string (static, not freed) */
uint32_t prop_id; /**< KMS property object ID */
unsigned int num_enum_values; /**< number of enum values */
struct drm_property_enum_info *enum_values; /**< array of enum values */
};
/**
* List of properties attached to DRM planes
*/
enum wdrm_plane_property {
WDRM_PLANE_TYPE = 0,
WDRM_PLANE_SRC_X,
WDRM_PLANE_SRC_Y,
WDRM_PLANE_SRC_W,
WDRM_PLANE_SRC_H,
WDRM_PLANE_CRTC_X,
WDRM_PLANE_CRTC_Y,
WDRM_PLANE_CRTC_W,
WDRM_PLANE_CRTC_H,
WDRM_PLANE_FB_ID,
WDRM_PLANE_CRTC_ID,
WDRM_PLANE_IN_FORMATS,
WDRM_PLANE__COUNT
};
/**
* Possible values for the WDRM_PLANE_TYPE property.
*/
enum wdrm_plane_type {
WDRM_PLANE_TYPE_PRIMARY = 0,
WDRM_PLANE_TYPE_CURSOR,
WDRM_PLANE_TYPE_OVERLAY,
WDRM_PLANE_TYPE__COUNT
};
static struct drm_property_enum_info plane_type_enums[] = {
[WDRM_PLANE_TYPE_PRIMARY] = {
.name = "Primary",
},
[WDRM_PLANE_TYPE_OVERLAY] = {
.name = "Overlay",
},
[WDRM_PLANE_TYPE_CURSOR] = {
.name = "Cursor",
},
};
static const struct drm_property_info plane_props[] = {
[WDRM_PLANE_TYPE] = {
.name = "type",
.enum_values = plane_type_enums,
.num_enum_values = WDRM_PLANE_TYPE__COUNT,
},
[WDRM_PLANE_SRC_X] = { .name = "SRC_X", },
[WDRM_PLANE_SRC_Y] = { .name = "SRC_Y", },
[WDRM_PLANE_SRC_W] = { .name = "SRC_W", },
[WDRM_PLANE_SRC_H] = { .name = "SRC_H", },
[WDRM_PLANE_CRTC_X] = { .name = "CRTC_X", },
[WDRM_PLANE_CRTC_Y] = { .name = "CRTC_Y", },
[WDRM_PLANE_CRTC_W] = { .name = "CRTC_W", },
[WDRM_PLANE_CRTC_H] = { .name = "CRTC_H", },
[WDRM_PLANE_FB_ID] = { .name = "FB_ID", },
[WDRM_PLANE_CRTC_ID] = { .name = "CRTC_ID", },
[WDRM_PLANE_IN_FORMATS] = { .name = "IN_FORMATS" },
};
/**
* List of properties attached to a DRM connector
*/
enum wdrm_connector_property {
WDRM_CONNECTOR_EDID = 0,
WDRM_CONNECTOR_DPMS,
WDRM_CONNECTOR_CRTC_ID,
WDRM_CONNECTOR_HDR10_METADATA,
WDRM_CONNECTOR__COUNT
};
enum wdrm_dpms_state {
WDRM_DPMS_STATE_OFF = 0,
WDRM_DPMS_STATE_ON,
WDRM_DPMS_STATE_STANDBY, /* unused */
WDRM_DPMS_STATE_SUSPEND, /* unused */
WDRM_DPMS_STATE__COUNT
};
static struct drm_property_enum_info dpms_state_enums[] = {
[WDRM_DPMS_STATE_OFF] = {
.name = "Off",
},
[WDRM_DPMS_STATE_ON] = {
.name = "On",
},
[WDRM_DPMS_STATE_STANDBY] = {
.name = "Standby",
},
[WDRM_DPMS_STATE_SUSPEND] = {
.name = "Suspend",
},
};
static const struct drm_property_info connector_props[] = {
[WDRM_CONNECTOR_EDID] = { .name = "EDID" },
[WDRM_CONNECTOR_DPMS] = {
.name = "DPMS",
.enum_values = dpms_state_enums,
.num_enum_values = WDRM_DPMS_STATE__COUNT,
},
[WDRM_CONNECTOR_CRTC_ID] = { .name = "CRTC_ID", },
[WDRM_CONNECTOR_HDR10_METADATA] = { .name = "HDR_SOURCE_METADATA", },
};
/**
* List of properties attached to DRM CRTCs
*/
enum wdrm_crtc_property {
WDRM_CRTC_MODE_ID = 0,
WDRM_CRTC_ACTIVE,
WDRM_CRTC__COUNT
};
static const struct drm_property_info crtc_props[] = {
[WDRM_CRTC_MODE_ID] = { .name = "MODE_ID", },
[WDRM_CRTC_ACTIVE] = { .name = "ACTIVE", },
};
/**
* Mode for drm_output_state_duplicate.
*/
enum drm_output_state_duplicate_mode {
DRM_OUTPUT_STATE_CLEAR_PLANES, /**< reset all planes to off */
DRM_OUTPUT_STATE_PRESERVE_PLANES, /**< preserve plane state */
};
/**
* Mode for drm_pending_state_apply and co.
*/
enum drm_state_apply_mode {
DRM_STATE_APPLY_SYNC, /**< state fully processed */
DRM_STATE_APPLY_ASYNC, /**< state pending event delivery */
DRM_STATE_TEST_ONLY, /**< test if the state can be applied */
};
struct drm_backend {
struct weston_backend base;
struct weston_compositor *compositor;
struct udev *udev;
struct wl_event_source *drm_source;
struct udev_monitor *udev_monitor;
struct wl_event_source *udev_drm_source;
struct {
int id;
int fd;
char *filename;
} drm;
struct gbm_device *gbm;
struct wl_listener session_listener;
uint32_t gbm_format;
/* hdr10 metadata blob id */
unsigned int hdr_blob_id;
/* we need these parameters in order to not fail drmModeAddFB2()
* due to out of bounds dimensions, and then mistakenly set
* sprites_are_broken:
*/
int min_width, max_width;
int min_height, max_height;
struct wl_list plane_list;
int sprites_are_broken;
int sprites_hidden;
void *repaint_data;
bool state_invalid;
/* CRTC IDs not used by any enabled output. */
struct wl_array unused_crtcs;
int cursors_are_broken;
bool universal_planes;
bool atomic_modeset;
int use_pixman;
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
int use_g2d;
#endif
bool use_pixman_shadow;
struct udev_input input;
int32_t cursor_width;
int32_t cursor_height;
uint32_t pageflip_timeout;
bool shutting_down;
bool aspect_ratio_supported;
};
struct drm_mode {
struct weston_mode base;
drmModeModeInfo mode_info;
uint32_t blob_id;
};
enum drm_fb_type {
BUFFER_INVALID = 0, /**< never used */
BUFFER_CLIENT, /**< directly sourced from client */
BUFFER_DMABUF, /**< imported from linux_dmabuf client */
BUFFER_PIXMAN_DUMB, /**< internal Pixman rendering */
BUFFER_GBM_SURFACE, /**< internal EGL rendering */
BUFFER_CURSOR, /**< internal cursor buffer */
};
struct drm_fb {
enum drm_fb_type type;
int refcnt;
uint32_t fb_id, size;
uint32_t handles[4];
uint32_t strides[4];
uint32_t offsets[4];
const struct pixel_format_info *format;
uint64_t modifier;
int width, height;
int fd;
struct weston_buffer_reference buffer_ref;
/* Used by gbm fbs */
struct gbm_bo *bo;
struct gbm_surface *gbm_surface;
/* Used by dumb fbs */
void *map;
};
struct drm_edid {
char eisa_id[13];
char monitor_name[13];
char pnp_id[5];
char serial_number[13];
};
/**
* Pending state holds one or more drm_output_state structures, collected from
* performing repaint. This pending state is transient, and only lives between
* beginning a repaint group and flushing the results: after flush, each
* output state will complete and be retired separately.
*/
struct drm_pending_state {
struct drm_backend *backend;
struct wl_list output_list;
};
/*
* Output state holds the dynamic state for one Weston output, i.e. a KMS CRTC,
* plus >= 1 each of encoder/connector/plane. Since everything but the planes
* is currently statically assigned per-output, we mainly use this to track
* plane state.
*
* pending_state is set when the output state is owned by a pending_state,
* i.e. when it is being constructed and has not yet been applied. When the
* output state has been applied, the owning pending_state is freed.
*/
struct drm_output_state {
struct drm_pending_state *pending_state;
struct drm_output *output;
struct wl_list link;
enum dpms_enum dpms;
struct wl_list plane_list;
};
/**
* Plane state holds the dynamic state for a plane: where it is positioned,
* and which buffer it is currently displaying.
*
* The plane state is owned by an output state, except when setting an initial
* state. See drm_output_state for notes on state object lifetime.
*/
struct drm_plane_state {
struct drm_plane *plane;
struct drm_output *output;
struct drm_output_state *output_state;
struct drm_fb *fb;
struct weston_view *ev; /**< maintained for drm_assign_planes only */
int32_t src_x, src_y;
uint32_t src_w, src_h;
int32_t dest_x, dest_y;
uint32_t dest_w, dest_h;
bool complete;
struct wl_list link; /* drm_output_state::plane_list */
};
/**
* A plane represents one buffer, positioned within a CRTC, and stacked
* relative to other planes on the same CRTC.
*
* Each CRTC has a 'primary plane', which use used to display the classic
* framebuffer contents, as accessed through the legacy drmModeSetCrtc
* call (which combines setting the CRTC's actual physical mode, and the
* properties of the primary plane).
*
* The cursor plane also has its own alternate legacy API.
*
* Other planes are used opportunistically to display content we do not
* wish to blit into the primary plane. These non-primary/cursor planes
* are referred to as 'sprites'.
*/
struct drm_plane {
struct weston_plane base;
struct drm_backend *backend;
enum wdrm_plane_type type;
uint32_t possible_crtcs;
uint32_t plane_id;
uint32_t count_formats;
struct drm_property_info props[WDRM_PLANE__COUNT];
/* The last state submitted to the kernel for this plane. */
struct drm_plane_state *state_cur;
struct wl_list link;
struct {
uint32_t format;
uint32_t count_modifiers;
uint64_t *modifiers;
} formats[];
};
struct drm_head {
struct weston_head base;
struct drm_backend *backend;
drmModeConnector *connector;
uint32_t connector_id;
struct drm_edid edid;
/* Holds the properties for the connector */
struct drm_property_info props_conn[WDRM_CONNECTOR__COUNT];
struct backlight *backlight;
drmModeModeInfo inherited_mode; /**< Original mode on the connector */
uint32_t inherited_crtc_id; /**< Original CRTC assignment */
};
struct drm_output {
struct weston_output base;
uint32_t crtc_id; /* object ID to pass to DRM functions */
int pipe; /* index of CRTC in resource array / bitmasks */
/* Holds the properties for the CRTC */
struct drm_property_info props_crtc[WDRM_CRTC__COUNT];
int vblank_pending;
int page_flip_pending;
int atomic_complete_pending;
int destroy_pending;
int disable_pending;
int dpms_off_pending;
struct drm_fb *gbm_cursor_fb[2];
struct drm_plane *cursor_plane;
struct weston_view *cursor_view;
int current_cursor;
struct gbm_surface *gbm_surface;
uint32_t gbm_format;
/* Plane being displayed directly on the CRTC */
struct drm_plane *scanout_plane;
/* The last state submitted to the kernel for this CRTC. */
struct drm_output_state *state_cur;
/* The previously-submitted state, where the hardware has not
* yet acknowledged completion of state_cur. */
struct drm_output_state *state_last;
struct drm_fb *dumb[2];
pixman_image_t *image[2];
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
struct g2d_surfaceEx g2d_image[2];
#endif
int current_image;
pixman_region32_t previous_damage;
struct vaapi_recorder *recorder;
struct wl_listener recorder_frame_listener;
struct wl_event_source *pageflip_timer;
};
static const char *const aspect_ratio_as_string[] = {
[WESTON_MODE_PIC_AR_NONE] = "",
[WESTON_MODE_PIC_AR_4_3] = " 4:3",
[WESTON_MODE_PIC_AR_16_9] = " 16:9",
[WESTON_MODE_PIC_AR_64_27] = " 64:27",
[WESTON_MODE_PIC_AR_256_135] = " 256:135",
};
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_OPENGL)
static struct gl_renderer_interface *gl_renderer;
#endif
#if defined(ENABLE_IMXG2D)
static struct g2d_renderer_interface *g2d_renderer;
#endif
#endif
static const char default_seat[] = "seat0";
static void
wl_array_remove_uint32(struct wl_array *array, uint32_t elm)
{
uint32_t *pos, *end;
end = (uint32_t *) ((char *) array->data + array->size);
wl_array_for_each(pos, array) {
if (*pos != elm)
continue;
array->size -= sizeof(*pos);
if (pos + 1 == end)
break;
memmove(pos, pos + 1, (char *) end - (char *) (pos + 1));
break;
}
}
static inline struct drm_head *
to_drm_head(struct weston_head *base)
{
return container_of(base, struct drm_head, base);
}
static inline struct drm_output *
to_drm_output(struct weston_output *base)
{
return container_of(base, struct drm_output, base);
}
static inline struct drm_backend *
to_drm_backend(struct weston_compositor *base)
{
return container_of(base->backend, struct drm_backend, base);
}
static int
pageflip_timeout(void *data) {
/*
* Our timer just went off, that means we're not receiving drm
* page flip events anymore for that output. Let's gracefully exit
* weston with a return value so devs can debug what's going on.
*/
struct drm_output *output = data;
struct weston_compositor *compositor = output->base.compositor;
weston_log("Pageflip timeout reached on output %s, your "
"driver is probably buggy! Exiting.\n",
output->base.name);
weston_compositor_exit_with_code(compositor, EXIT_FAILURE);
return 0;
}
/* Creates the pageflip timer. Note that it isn't armed by default */
static int
drm_output_pageflip_timer_create(struct drm_output *output)
{
struct wl_event_loop *loop = NULL;
struct weston_compositor *ec = output->base.compositor;
loop = wl_display_get_event_loop(ec->wl_display);
assert(loop);
output->pageflip_timer = wl_event_loop_add_timer(loop,
pageflip_timeout,
output);
if (output->pageflip_timer == NULL) {
weston_log("creating drm pageflip timer failed: %m\n");
return -1;
}
return 0;
}
static inline struct drm_mode *
to_drm_mode(struct weston_mode *base)
{
return container_of(base, struct drm_mode, base);
}
/**
* Get the current value of a KMS property
*
* Given a drmModeObjectGetProperties return, as well as the drm_property_info
* for the target property, return the current value of that property,
* with an optional default. If the property is a KMS enum type, the return
* value will be translated into the appropriate internal enum.
*
* If the property is not present, the default value will be returned.
*
* @param info Internal structure for property to look up
* @param props Raw KMS properties for the target object
* @param def Value to return if property is not found
*/
static uint64_t
drm_property_get_value(struct drm_property_info *info,
const drmModeObjectProperties *props,
uint64_t def)
{
unsigned int i;
if (info->prop_id == 0)
return def;
for (i = 0; i < props->count_props; i++) {
unsigned int j;
if (props->props[i] != info->prop_id)
continue;
/* Simple (non-enum) types can return the value directly */
if (info->num_enum_values == 0)
return props->prop_values[i];
/* Map from raw value to enum value */
for (j = 0; j < info->num_enum_values; j++) {
if (!info->enum_values[j].valid)
continue;
if (info->enum_values[j].value != props->prop_values[i])
continue;
return j;
}
/* We don't have a mapping for this enum; return default. */
break;
}
return def;
}
/**
* Cache DRM property values
*
* Update a per-object array of drm_property_info structures, given the
* DRM properties of the object.
*
* Call this every time an object newly appears (note that only connectors
* can be hotplugged), the first time it is seen, or when its status changes
* in a way which invalidates the potential property values (currently, the
* only case for this is connector hotplug).
*
* This updates the property IDs and enum values within the drm_property_info
* array.
*
* DRM property enum values are dynamic at runtime; the user must query the
* property to find out the desired runtime value for a requested string
* name. Using the 'type' field on planes as an example, there is no single
* hardcoded constant for primary plane types; instead, the property must be
* queried at runtime to find the value associated with the string "Primary".
*
* This helper queries and caches the enum values, to allow us to use a set
* of compile-time-constant enums portably across various implementations.
* The values given in enum_names are searched for, and stored in the
* same-indexed field of the map array.
*
* @param b DRM backend object
* @param src DRM property info array to source from
* @param info DRM property info array to copy into
* @param num_infos Number of entries in the source array
* @param props DRM object properties for the object
*/
static void
drm_property_info_populate(struct drm_backend *b,
const struct drm_property_info *src,
struct drm_property_info *info,
unsigned int num_infos,
drmModeObjectProperties *props)
{
drmModePropertyRes *prop;
unsigned i, j;
for (i = 0; i < num_infos; i++) {
unsigned int j;
info[i].name = src[i].name;
info[i].prop_id = 0;
info[i].num_enum_values = src[i].num_enum_values;
if (src[i].num_enum_values == 0)
continue;
info[i].enum_values =
malloc(src[i].num_enum_values *
sizeof(*info[i].enum_values));
assert(info[i].enum_values);
for (j = 0; j < info[i].num_enum_values; j++) {
info[i].enum_values[j].name = src[i].enum_values[j].name;
info[i].enum_values[j].valid = false;
}
}
for (i = 0; i < props->count_props; i++) {
unsigned int k;
prop = drmModeGetProperty(b->drm.fd, props->props[i]);
if (!prop)
continue;
for (j = 0; j < num_infos; j++) {
if (!strcmp(prop->name, info[j].name))
break;
}
/* We don't know/care about this property. */
if (j == num_infos) {
#ifdef DEBUG
weston_log("DRM debug: unrecognized property %u '%s'\n",
prop->prop_id, prop->name);
#endif
drmModeFreeProperty(prop);
continue;
}
if (info[j].num_enum_values == 0 &&
(prop->flags & DRM_MODE_PROP_ENUM)) {
weston_log("DRM: expected property %s to not be an"
" enum, but it is; ignoring\n", prop->name);
drmModeFreeProperty(prop);
continue;
}
info[j].prop_id = props->props[i];
if (info[j].num_enum_values == 0) {
drmModeFreeProperty(prop);
continue;
}
if (!(prop->flags & DRM_MODE_PROP_ENUM)) {
weston_log("DRM: expected property %s to be an enum,"
" but it is not; ignoring\n", prop->name);
drmModeFreeProperty(prop);
info[j].prop_id = 0;
continue;
}
for (k = 0; k < info[j].num_enum_values; k++) {
int l;
for (l = 0; l < prop->count_enums; l++) {
if (!strcmp(prop->enums[l].name,
info[j].enum_values[k].name))
break;
}
if (l == prop->count_enums)
continue;
info[j].enum_values[k].valid = true;
info[j].enum_values[k].value = prop->enums[l].value;
}
drmModeFreeProperty(prop);
}
#ifdef DEBUG
for (i = 0; i < num_infos; i++) {
if (info[i].prop_id == 0)
weston_log("DRM warning: property '%s' missing\n",
info[i].name);
}
#endif
}
/**
* Free DRM property information
*
* Frees all memory associated with a DRM property info array and zeroes
* it out, leaving it usable for a further drm_property_info_update() or
* drm_property_info_free().
*
* @param info DRM property info array
* @param num_props Number of entries in array to free
*/
static void
drm_property_info_free(struct drm_property_info *info, int num_props)
{
int i;
for (i = 0; i < num_props; i++)
free(info[i].enum_values);
memset(info, 0, sizeof(*info) * num_props);
}
static void
drm_output_set_cursor(struct drm_output_state *output_state);
static void
drm_output_update_msc(struct drm_output *output, unsigned int seq);
static void
drm_output_destroy(struct weston_output *output_base);
/**
* Returns true if the plane can be used on the given output for its current
* repaint cycle.
*/
static bool
drm_plane_is_available(struct drm_plane *plane, struct drm_output *output)
{
assert(plane->state_cur);
/* The plane still has a request not yet completed by the kernel. */
if (!plane->state_cur->complete)
return false;
/* The plane is still active on another output. */
if (plane->state_cur->output && plane->state_cur->output != output)
return false;
/* Check whether the plane can be used with this CRTC; possible_crtcs
* is a bitmask of CRTC indices (pipe), rather than CRTC object ID. */
return !!(plane->possible_crtcs & (1 << output->pipe));
}
static struct drm_output *
drm_output_find_by_crtc(struct drm_backend *b, uint32_t crtc_id)
{
struct drm_output *output;
wl_list_for_each(output, &b->compositor->output_list, base.link) {
if (output->crtc_id == crtc_id)
return output;
}
return NULL;
}
static struct drm_head *
drm_head_find_by_connector(struct drm_backend *backend, uint32_t connector_id)
{
struct weston_head *base;
struct drm_head *head;
wl_list_for_each(base,
&backend->compositor->head_list, compositor_link) {
head = to_drm_head(base);
if (head->connector_id == connector_id)
return head;
}
return NULL;
}
static void
drm_fb_destroy(struct drm_fb *fb)
{
if (fb->fb_id != 0)
drmModeRmFB(fb->fd, fb->fb_id);
weston_buffer_reference(&fb->buffer_ref, NULL);
free(fb);
}
static void
drm_fb_destroy_dumb(struct drm_fb *fb)
{
struct drm_mode_destroy_dumb destroy_arg;
assert(fb->type == BUFFER_PIXMAN_DUMB);
if (fb->map && fb->size > 0)
munmap(fb->map, fb->size);
memset(&destroy_arg, 0, sizeof(destroy_arg));
destroy_arg.handle = fb->handles[0];
drmIoctl(fb->fd, DRM_IOCTL_MODE_DESTROY_DUMB, &destroy_arg);
drm_fb_destroy(fb);
}
static void
drm_fb_destroy_gbm(struct gbm_bo *bo, void *data)
{
struct drm_fb *fb = data;
assert(fb->type == BUFFER_GBM_SURFACE || fb->type == BUFFER_CLIENT ||
fb->type == BUFFER_CURSOR);
drm_fb_destroy(fb);
}
static int
drm_fb_addfb(struct drm_fb *fb)
{
int ret = -EINVAL;
#ifdef HAVE_DRM_ADDFB2_MODIFIERS
uint64_t mods[4] = { };
int width, height;
size_t i;
#endif
/* If we have a modifier set, we must only use the WithModifiers
* entrypoint; we cannot import it through legacy ioctls. */
if (fb->modifier != DRM_FORMAT_MOD_INVALID) {
/* KMS demands that if a modifier is set, it must be the same
* for all planes. */
#ifdef HAVE_DRM_ADDFB2_MODIFIERS
for (i = 0; i < ARRAY_LENGTH(mods) && fb->handles[i]; i++)
mods[i] = fb->modifier;
if (fb->modifier == DRM_FORMAT_MOD_AMPHION_TILED) {
width = ALIGNTO (fb->width, 8);
height = ALIGNTO (fb->height, 256);
} else {
width = fb->width;
height = fb->height;
}
ret = drmModeAddFB2WithModifiers(fb->fd, width, height,
fb->format->format,
fb->handles, fb->strides,
fb->offsets, mods, &fb->fb_id,
DRM_MODE_FB_MODIFIERS);
#endif
return ret;
}
ret = drmModeAddFB2(fb->fd, fb->width, fb->height, fb->format->format,
fb->handles, fb->strides, fb->offsets, &fb->fb_id,
0);
if (ret == 0)
return 0;
/* Legacy AddFB can't always infer the format from depth/bpp alone, so
* check if our format is one of the lucky ones. */
if (!fb->format->depth || !fb->format->bpp)
return ret;
/* Cannot fall back to AddFB for multi-planar formats either. */
if (fb->handles[1] || fb->handles[2] || fb->handles[3])
return ret;
ret = drmModeAddFB(fb->fd, fb->width, fb->height,
fb->format->depth, fb->format->bpp,
fb->strides[0], fb->handles[0], &fb->fb_id);
return ret;
}
static struct drm_fb *
drm_fb_create_dumb(struct drm_backend *b, int width, int height,
uint32_t format)
{
struct drm_fb *fb;
int ret;
struct drm_mode_create_dumb create_arg;
struct drm_mode_destroy_dumb destroy_arg;
struct drm_mode_map_dumb map_arg;
fb = zalloc(sizeof *fb);
if (!fb)
return NULL;
fb->refcnt = 1;
fb->format = pixel_format_get_info(format);
if (!fb->format) {
weston_log("failed to look up format 0x%lx\n",
(unsigned long) format);
goto err_fb;
}
if (!fb->format->depth || !fb->format->bpp) {
weston_log("format 0x%lx is not compatible with dumb buffers\n",
(unsigned long) format);
goto err_fb;
}
memset(&create_arg, 0, sizeof create_arg);
create_arg.bpp = fb->format->bpp;
create_arg.width = width;
create_arg.height = height;
ret = drmIoctl(b->drm.fd, DRM_IOCTL_MODE_CREATE_DUMB, &create_arg);
if (ret)
goto err_fb;
fb->type = BUFFER_PIXMAN_DUMB;
fb->modifier = DRM_FORMAT_MOD_INVALID;
fb->handles[0] = create_arg.handle;
fb->strides[0] = create_arg.pitch;
fb->size = create_arg.size;
fb->width = width;
fb->height = height;
fb->fd = b->drm.fd;
if (drm_fb_addfb(fb) != 0) {
weston_log("failed to create kms fb: %m\n");
goto err_bo;
}
memset(&map_arg, 0, sizeof map_arg);
map_arg.handle = fb->handles[0];
ret = drmIoctl(fb->fd, DRM_IOCTL_MODE_MAP_DUMB, &map_arg);
if (ret)
goto err_add_fb;
fb->map = mmap(NULL, fb->size, PROT_WRITE,
MAP_SHARED, b->drm.fd, map_arg.offset);
if (fb->map == MAP_FAILED)
goto err_add_fb;
return fb;
err_add_fb:
drmModeRmFB(b->drm.fd, fb->fb_id);
err_bo:
memset(&destroy_arg, 0, sizeof(destroy_arg));
destroy_arg.handle = create_arg.handle;
drmIoctl(b->drm.fd, DRM_IOCTL_MODE_DESTROY_DUMB, &destroy_arg);
err_fb:
free(fb);
return NULL;
}
static struct drm_fb *
drm_fb_ref(struct drm_fb *fb)
{
fb->refcnt++;
return fb;
}
static void
drm_fb_destroy_dmabuf(struct drm_fb *fb)
{
/* We deliberately do not close the GEM handles here; GBM manages
* their lifetime through the BO. */
if (fb->bo)
gbm_bo_destroy(fb->bo);
drm_fb_destroy(fb);
}
static struct drm_fb *
drm_fb_get_from_dmabuf(struct linux_dmabuf_buffer *dmabuf,
struct drm_backend *backend, bool is_opaque)
{
#ifdef HAVE_GBM_FD_IMPORT
struct drm_fb *fb;
struct gbm_import_fd_data import_legacy = {
.width = dmabuf->attributes.width,
.height = dmabuf->attributes.height,
.format = dmabuf->attributes.format,
.stride = dmabuf->attributes.stride[0],
.fd = dmabuf->attributes.fd[0],
};
struct gbm_import_fd_modifier_data import_mod = {
.width = dmabuf->attributes.width,
.height = dmabuf->attributes.height,
.format = dmabuf->attributes.format,
.num_fds = dmabuf->attributes.n_planes,
.modifier = dmabuf->attributes.modifier[0],
};
int i;
uint32_t gem_handle[MAX_DMABUF_PLANES] = {0};
/* XXX: TODO:
*
* Currently the buffer is rejected if any dmabuf attribute
* flag is set. This keeps us from passing an inverted /
* interlaced / bottom-first buffer (or any other type that may
* be added in the future) through to an overlay. Ultimately,
* these types of buffers should be handled through buffer
* transforms and not as spot-checks requiring specific
* knowledge. */
if (dmabuf->attributes.flags)
return NULL;
fb = zalloc(sizeof *fb);
if (fb == NULL)
return NULL;
fb->refcnt = 1;
fb->type = BUFFER_DMABUF;
fb->width = dmabuf->attributes.width;
fb->height = dmabuf->attributes.height;
fb->modifier = dmabuf->attributes.modifier[0];
fb->size = 0;
fb->fd = backend->drm.fd;
static_assert(ARRAY_LENGTH(fb->strides) ==
ARRAY_LENGTH(dmabuf->attributes.stride),
"drm_fb and dmabuf stride size must match");
static_assert(sizeof(fb->strides) == sizeof(dmabuf->attributes.stride),
"drm_fb and dmabuf stride size must match");
memcpy(fb->strides, dmabuf->attributes.stride, sizeof(fb->strides));
static_assert(ARRAY_LENGTH(fb->offsets) ==
ARRAY_LENGTH(dmabuf->attributes.offset),
"drm_fb and dmabuf offset size must match");
static_assert(sizeof(fb->offsets) == sizeof(dmabuf->attributes.offset),
"drm_fb and dmabuf offset size must match");
memcpy(fb->offsets, dmabuf->attributes.offset, sizeof(fb->offsets));
fb->format = pixel_format_get_info(dmabuf->attributes.format);
if (!fb->format) {
weston_log("couldn't look up format info for 0x%lx\n",
(unsigned long) dmabuf->attributes.format);
goto err_free;
}
if (is_opaque)
fb->format = pixel_format_get_opaque_substitute(fb->format);
if (backend->min_width > fb->width ||
fb->width > backend->max_width ||
backend->min_height > fb->height ||
fb->height > backend->max_height) {
weston_log("bo geometry out of bounds\n");
goto err_free;
}
for (i = 0; i < dmabuf->attributes.n_planes; i++) {
int ret;
ret = drmPrimeFDToHandle (fb->fd, dmabuf->attributes.fd[i], &gem_handle[i]);
if (ret) {
weston_log ("got gem_handle %x\n", gem_handle[i]);
goto err_free;
}
fb->handles[i] = gem_handle[i];
}
if (gem_handle[0] != 0)
goto add_fb;
static_assert(ARRAY_LENGTH(import_mod.fds) ==
ARRAY_LENGTH(dmabuf->attributes.fd),
"GBM and linux_dmabuf FD size must match");
static_assert(sizeof(import_mod.fds) == sizeof(dmabuf->attributes.fd),
"GBM and linux_dmabuf FD size must match");
memcpy(import_mod.fds, dmabuf->attributes.fd, sizeof(import_mod.fds));
static_assert(ARRAY_LENGTH(import_mod.strides) ==
ARRAY_LENGTH(dmabuf->attributes.stride),
"GBM and linux_dmabuf stride size must match");
static_assert(sizeof(import_mod.strides) ==
sizeof(dmabuf->attributes.stride),
"GBM and linux_dmabuf stride size must match");
memcpy(import_mod.strides, dmabuf->attributes.stride,
sizeof(import_mod.strides));
static_assert(ARRAY_LENGTH(import_mod.offsets) ==
ARRAY_LENGTH(dmabuf->attributes.offset),
"GBM and linux_dmabuf offset size must match");
static_assert(sizeof(import_mod.offsets) ==
sizeof(dmabuf->attributes.offset),
"GBM and linux_dmabuf offset size must match");
memcpy(import_mod.offsets, dmabuf->attributes.offset,
sizeof(import_mod.offsets));
/* The legacy FD-import path does not allow us to supply modifiers,
* multiple planes, or buffer offsets. */
if (dmabuf->attributes.modifier[0] != DRM_FORMAT_MOD_INVALID ||
import_mod.num_fds > 1 ||
import_mod.offsets[0] > 0) {
fb->bo = gbm_bo_import(backend->gbm, GBM_BO_IMPORT_FD_MODIFIER,
&import_mod,
GBM_BO_USE_SCANOUT);
} else {
fb->bo = gbm_bo_import(backend->gbm, GBM_BO_IMPORT_FD,
&import_legacy,
GBM_BO_USE_SCANOUT);
}
if (!fb->bo)
goto err_free;
for (i = 0; i < dmabuf->attributes.n_planes; i++) {
fb->handles[i] = gbm_bo_get_handle_for_plane(fb->bo, i).u32;
if (!fb->handles[i])
goto err_free;
}
add_fb:
if (drm_fb_addfb(fb) != 0)
goto err_free;
if (gem_handle[0] != 0) {
for (i = 0; i < dmabuf->attributes.n_planes; i++) {
struct drm_gem_close arg = { gem_handle[i], };
drmIoctl (fb->fd, DRM_IOCTL_GEM_CLOSE, &arg);
}
}
return fb;
err_free:
drm_fb_destroy_dmabuf(fb);
#endif
return NULL;
}
static struct drm_fb *
drm_fb_get_from_bo(struct gbm_bo *bo, struct drm_backend *backend,
bool is_opaque, enum drm_fb_type type)
{
struct drm_fb *fb = gbm_bo_get_user_data(bo);
#ifdef HAVE_GBM_MODIFIERS
int i;
#endif
if (fb) {
assert(fb->type == type);
return drm_fb_ref(fb);
}
fb = zalloc(sizeof *fb);
if (fb == NULL)
return NULL;
fb->type = type;
fb->refcnt = 1;
fb->bo = bo;
fb->fd = backend->drm.fd;
fb->width = gbm_bo_get_width(bo);
fb->height = gbm_bo_get_height(bo);
fb->format = pixel_format_get_info(gbm_bo_get_format(bo));
fb->size = 0;
#ifdef HAVE_GBM_MODIFIERS
fb->modifier = gbm_bo_get_modifier(bo);
for (i = 0; i < gbm_bo_get_plane_count(bo); i++) {
fb->strides[i] = gbm_bo_get_stride_for_plane(bo, i);
fb->handles[i] = gbm_bo_get_handle_for_plane(bo, i).u32;
fb->offsets[i] = gbm_bo_get_offset(bo, i);
}
#else
fb->strides[0] = gbm_bo_get_stride(bo);
fb->handles[0] = gbm_bo_get_handle(bo).u32;
fb->modifier = DRM_FORMAT_MOD_INVALID;
#endif
if (!fb->format) {
weston_log("couldn't look up format 0x%lx\n",
(unsigned long) gbm_bo_get_format(bo));
goto err_free;
}
if (backend->min_width > fb->width ||
fb->width > backend->max_width ||
backend->min_height > fb->height ||
fb->height > backend->max_height) {
weston_log("bo geometry out of bounds\n");
goto err_free;
}
if (drm_fb_addfb(fb) != 0) {
if (type == BUFFER_GBM_SURFACE)
weston_log("failed to create kms fb: %m\n");
goto err_free;
}
gbm_bo_set_user_data(bo, fb, drm_fb_destroy_gbm);
return fb;
err_free:
free(fb);
return NULL;
}
static void
drm_fb_set_buffer(struct drm_fb *fb, struct weston_buffer *buffer)
{
assert(fb->buffer_ref.buffer == NULL);
assert(fb->type == BUFFER_CLIENT || fb->type == BUFFER_DMABUF);
weston_buffer_reference(&fb->buffer_ref, buffer);
}
static void
drm_fb_unref(struct drm_fb *fb)
{
if (!fb)
return;
assert(fb->refcnt > 0);
if (--fb->refcnt > 0)
return;
switch (fb->type) {
case BUFFER_PIXMAN_DUMB:
drm_fb_destroy_dumb(fb);
break;
case BUFFER_CURSOR:
case BUFFER_CLIENT:
gbm_bo_destroy(fb->bo);
break;
case BUFFER_GBM_SURFACE:
gbm_surface_release_buffer(fb->gbm_surface, fb->bo);
break;
case BUFFER_DMABUF:
drm_fb_destroy_dmabuf(fb);
break;
default:
assert(NULL);
break;
}
}
/**
* Allocate a new, empty, plane state.
*/
static struct drm_plane_state *
drm_plane_state_alloc(struct drm_output_state *state_output,
struct drm_plane *plane)
{
struct drm_plane_state *state = zalloc(sizeof(*state));
assert(state);
state->output_state = state_output;
state->plane = plane;
/* Here we only add the plane state to the desired link, and not
* set the member. Having an output pointer set means that the
* plane will be displayed on the output; this won't be the case
* when we go to disable a plane. In this case, it must be part of
* the commit (and thus the output state), but the member must be
* NULL, as it will not be on any output when the state takes
* effect.
*/
if (state_output)
wl_list_insert(&state_output->plane_list, &state->link);
else
wl_list_init(&state->link);
return state;
}
/**
* Free an existing plane state. As a special case, the state will not
* normally be freed if it is the current state; see drm_plane_set_state.
*/
static void
drm_plane_state_free(struct drm_plane_state *state, bool force)
{
if (!state)
return;
wl_list_remove(&state->link);
wl_list_init(&state->link);
state->output_state = NULL;
if (force || state != state->plane->state_cur) {
drm_fb_unref(state->fb);
free(state);
}
}
/**
* Duplicate an existing plane state into a new plane state, storing it within
* the given output state. If the output state already contains a plane state
* for the drm_plane referenced by 'src', that plane state is freed first.
*/
static struct drm_plane_state *
drm_plane_state_duplicate(struct drm_output_state *state_output,
struct drm_plane_state *src)
{
struct drm_plane_state *dst = malloc(sizeof(*dst));
struct drm_plane_state *old, *tmp;
assert(src);
assert(dst);
*dst = *src;
wl_list_init(&dst->link);
wl_list_for_each_safe(old, tmp, &state_output->plane_list, link) {
/* Duplicating a plane state into the same output state, so
* it can replace itself with an identical copy of itself,
* makes no sense. */
assert(old != src);
if (old->plane == dst->plane)
drm_plane_state_free(old, false);
}
wl_list_insert(&state_output->plane_list, &dst->link);
if (src->fb)
dst->fb = drm_fb_ref(src->fb);
dst->output_state = state_output;
dst->complete = false;
return dst;
}
/**
* Remove a plane state from an output state; if the plane was previously
* enabled, then replace it with a disabling state. This ensures that the
* output state was untouched from it was before the plane state was
* modified by the caller of this function.
*
* This is required as drm_output_state_get_plane may either allocate a
* new plane state, in which case this function will just perform a matching
* drm_plane_state_free, or it may instead repurpose an existing disabling
* state (if the plane was previously active), in which case this function
* will reset it.
*/
static void
drm_plane_state_put_back(struct drm_plane_state *state)
{
struct drm_output_state *state_output;
struct drm_plane *plane;
if (!state)
return;
state_output = state->output_state;
plane = state->plane;
drm_plane_state_free(state, false);
/* Plane was previously disabled; no need to keep this temporary
* state around. */
if (!plane->state_cur->fb)
return;
(void) drm_plane_state_alloc(state_output, plane);
}
static bool
drm_view_transform_supported(struct weston_view *ev, struct weston_output *output)
{
struct weston_buffer_viewport *viewport = &ev->surface->buffer_viewport;
/* This will incorrectly disallow cases where the combination of
* buffer and view transformations match the output transform.
* Fixing this requires a full analysis of the transformation
* chain. */
if (ev->transform.enabled &&
ev->transform.matrix.type >= WESTON_MATRIX_TRANSFORM_ROTATE)
return false;
if (viewport->buffer.transform != output->transform)
return false;
return true;
}
static void
drm_clip_overlay_coordinate(struct drm_output * output,
struct drm_plane_state * state)
{
uint32_t src_w = state->src_w >> 16;
uint32_t src_h = state->src_h >> 16;
/* handle out of screen case */
if (state->dest_x + state->dest_w > output->base.current_mode->width){
int32_t crop_width = output->base.current_mode->width - state->dest_x;
if(crop_width > 0) {
state->src_w = ALIGNTO((crop_width * src_w / state->dest_w), 2) << 16;
state->dest_w = crop_width;
}
}
if (state->dest_y + state->dest_h > output->base.current_mode->height){
int32_t crop_height = output->base.current_mode->height - state->dest_y;
if(crop_height > 0) {
state->src_h = ALIGNTO((crop_height * src_h / state->dest_h), 2) << 16;
state->dest_h = crop_height;
}
}
}
/**
* Given a weston_view, fill the drm_plane_state's co-ordinates to display on
* a given plane.
*/
static bool
drm_plane_state_coords_for_view(struct drm_plane_state *state,
struct weston_view *ev)
{
struct drm_output *output = state->output;
struct weston_buffer_viewport *viewport = &ev->surface->buffer_viewport;
pixman_region32_t dest_rect;
pixman_box32_t *box, tbox;
int32_t scale;
int src_x, src_y, src_width, src_height;
if (!drm_view_transform_supported(ev, &output->base))
return false;
if (viewport->buffer.scale != output->base.current_scale){
scale = MAX(viewport->buffer.scale, output->base.current_scale);
} else {
scale = output->base.current_scale;
}
/* Update the base weston_plane co-ordinates. */
box = pixman_region32_extents(&ev->transform.boundingbox);
state->plane->base.x = box->x1;
state->plane->base.y = box->y1;
/* First calculate the destination co-ordinates by taking the
* area of the view which is visible on this output, performing any
* transforms to account for output rotation and scale as necessary. */
pixman_region32_init(&dest_rect);
pixman_region32_intersect(&dest_rect, &ev->transform.boundingbox,
&output->base.region);
pixman_region32_translate(&dest_rect, -output->base.x, -output->base.y);
box = pixman_region32_extents(&dest_rect);
tbox = weston_transformed_rect(output->base.width,
output->base.height,
output->base.transform,
scale,
*box);
state->dest_x = tbox.x1;
state->dest_y = tbox.y1;
state->dest_w = tbox.x2 - tbox.x1;
state->dest_h = tbox.y2 - tbox.y1;
pixman_region32_fini(&dest_rect);
src_x = wl_fixed_to_int (viewport->buffer.src_x);
src_y = wl_fixed_to_int (viewport->buffer.src_y);
src_width = wl_fixed_to_int (viewport->buffer.src_width);
src_height = wl_fixed_to_int (viewport->buffer.src_height);
if(src_width != -1 && src_width > 0 && src_x >=0 && src_y >= 0
&& src_x * scale < state->fb->width
&& src_y * scale < state->fb->height)
{
state->src_x = (src_x * scale) << 16;
state->src_y = (src_y * scale) << 16;
state->src_w = MIN(state->fb->width - src_x * scale, src_width * scale) << 16;
state->src_h = MIN(state->fb->height - src_y * scale, src_height * scale) << 16;
}
else
{
state->src_x = 0;
state->src_y = 0;
state->src_w = state->fb->width << 16;
state->src_h = state->fb->height << 16;
}
drm_clip_overlay_coordinate(output, state);
return true;
}
static bool
drm_view_is_opaque(struct weston_view *ev)
{
pixman_region32_t r;
bool ret = false;
pixman_region32_init_rect(&r, 0, 0,
ev->surface->width,
ev->surface->height);
pixman_region32_subtract(&r, &r, &ev->surface->opaque);
if (!pixman_region32_not_empty(&r))
ret = true;
pixman_region32_fini(&r);
return ret;
}
static struct drm_fb *
drm_fb_get_from_view(struct drm_output_state *state, struct weston_view *ev)
{
struct drm_output *output = state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct weston_buffer *buffer = ev->surface->buffer_ref.buffer;
bool is_opaque = drm_view_is_opaque(ev);
struct linux_dmabuf_buffer *dmabuf;
struct drm_fb *fb;
if (ev->alpha != 1.0f)
return NULL;
if (!drm_view_transform_supported(ev, &output->base))
return NULL;
if (!buffer)
return NULL;
if (wl_shm_buffer_get(buffer->resource))
return NULL;
/* GBM is used for dmabuf import as well as from client wl_buffer. */
if (!b->gbm)
return NULL;
dmabuf = linux_dmabuf_buffer_get(buffer->resource);
if (dmabuf) {
fb = drm_fb_get_from_dmabuf(dmabuf, b, is_opaque);
if (!fb)
return NULL;
} else {
struct gbm_bo *bo;
bo = gbm_bo_import(b->gbm, GBM_BO_IMPORT_WL_BUFFER,
buffer->resource, GBM_BO_USE_SCANOUT);
if (!bo)
return NULL;
fb = drm_fb_get_from_bo(bo, b, is_opaque, BUFFER_CLIENT);
if (!fb) {
gbm_bo_destroy(bo);
return NULL;
}
}
drm_fb_set_buffer(fb, buffer);
return fb;
}
/**
* Return a plane state from a drm_output_state.
*/
static struct drm_plane_state *
drm_output_state_get_existing_plane(struct drm_output_state *state_output,
struct drm_plane *plane)
{
struct drm_plane_state *ps;
wl_list_for_each(ps, &state_output->plane_list, link) {
if (ps->plane == plane)
return ps;
}
return NULL;
}
/**
* Return a plane state from a drm_output_state, either existing or
* freshly allocated.
*/
static struct drm_plane_state *
drm_output_state_get_plane(struct drm_output_state *state_output,
struct drm_plane *plane)
{
struct drm_plane_state *ps;
ps = drm_output_state_get_existing_plane(state_output, plane);
if (ps)
return ps;
return drm_plane_state_alloc(state_output, plane);
}
/**
* Allocate a new, empty drm_output_state. This should not generally be used
* in the repaint cycle; see drm_output_state_duplicate.
*/
static struct drm_output_state *
drm_output_state_alloc(struct drm_output *output,
struct drm_pending_state *pending_state)
{
struct drm_output_state *state = zalloc(sizeof(*state));
assert(state);
state->output = output;
state->dpms = WESTON_DPMS_OFF;
state->pending_state = pending_state;
if (pending_state)
wl_list_insert(&pending_state->output_list, &state->link);
else
wl_list_init(&state->link);
wl_list_init(&state->plane_list);
return state;
}
/**
* Duplicate an existing drm_output_state into a new one. This is generally
* used during the repaint cycle, to capture the existing state of an output
* and modify it to create a new state to be used.
*
* The mode determines whether the output will be reset to an a blank state,
* or an exact mirror of the current state.
*/
static struct drm_output_state *
drm_output_state_duplicate(struct drm_output_state *src,
struct drm_pending_state *pending_state,
enum drm_output_state_duplicate_mode plane_mode)
{
struct drm_output_state *dst = malloc(sizeof(*dst));
struct drm_plane_state *ps;
assert(dst);
/* Copy the whole structure, then individually modify the
* pending_state, as well as the list link into our pending
* state. */
*dst = *src;
dst->pending_state = pending_state;
if (pending_state)
wl_list_insert(&pending_state->output_list, &dst->link);
else
wl_list_init(&dst->link);
wl_list_init(&dst->plane_list);
wl_list_for_each(ps, &src->plane_list, link) {
/* Don't carry planes which are now disabled; these should be
* free for other outputs to reuse. */
if (!ps->output)
continue;
if (plane_mode == DRM_OUTPUT_STATE_CLEAR_PLANES)
(void) drm_plane_state_alloc(dst, ps->plane);
else
(void) drm_plane_state_duplicate(dst, ps);
}
return dst;
}
/**
* Free an unused drm_output_state.
*/
static void
drm_output_state_free(struct drm_output_state *state)
{
struct drm_plane_state *ps, *next;
if (!state)
return;
wl_list_for_each_safe(ps, next, &state->plane_list, link)
drm_plane_state_free(ps, false);
wl_list_remove(&state->link);
free(state);
}
/**
* Get output state to disable output
*
* Returns a pointer to an output_state object which can be used to disable
* an output (e.g. DPMS off).
*
* @param pending_state The pending state object owning this update
* @param output The output to disable
* @returns A drm_output_state to disable the output
*/
static struct drm_output_state *
drm_output_get_disable_state(struct drm_pending_state *pending_state,
struct drm_output *output)
{
struct drm_output_state *output_state;
output_state = drm_output_state_duplicate(output->state_cur,
pending_state,
DRM_OUTPUT_STATE_CLEAR_PLANES);
output_state->dpms = WESTON_DPMS_OFF;
return output_state;
}
/**
* Allocate a new drm_pending_state
*
* Allocate a new, empty, 'pending state' structure to be used across a
* repaint cycle or similar.
*
* @param backend DRM backend
* @returns Newly-allocated pending state structure
*/
static struct drm_pending_state *
drm_pending_state_alloc(struct drm_backend *backend)
{
struct drm_pending_state *ret;
ret = calloc(1, sizeof(*ret));
if (!ret)
return NULL;
ret->backend = backend;
wl_list_init(&ret->output_list);
return ret;
}
/**
* Free a drm_pending_state structure
*
* Frees a pending_state structure, as well as any output_states connected
* to this pending state.
*
* @param pending_state Pending state structure to free
*/
static void
drm_pending_state_free(struct drm_pending_state *pending_state)
{
struct drm_output_state *output_state, *tmp;
if (!pending_state)
return;
wl_list_for_each_safe(output_state, tmp, &pending_state->output_list,
link) {
drm_output_state_free(output_state);
}
free(pending_state);
}
/**
* Find an output state in a pending state
*
* Given a pending_state structure, find the output_state for a particular
* output.
*
* @param pending_state Pending state structure to search
* @param output Output to find state for
* @returns Output state if present, or NULL if not
*/
static struct drm_output_state *
drm_pending_state_get_output(struct drm_pending_state *pending_state,
struct drm_output *output)
{
struct drm_output_state *output_state;
wl_list_for_each(output_state, &pending_state->output_list, link) {
if (output_state->output == output)
return output_state;
}
return NULL;
}
static int drm_pending_state_apply_sync(struct drm_pending_state *state);
static int drm_pending_state_test(struct drm_pending_state *state);
/**
* Mark a drm_output_state (the output's last state) as complete. This handles
* any post-completion actions such as updating the repaint timer, disabling the
* output, and finally freeing the state.
*/
static void
drm_output_update_complete(struct drm_output *output, uint32_t flags,
unsigned int sec, unsigned int usec)
{
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_plane_state *ps;
struct timespec ts;
/* Stop the pageflip timer instead of rearming it here */
if (output->pageflip_timer)
wl_event_source_timer_update(output->pageflip_timer, 0);
wl_list_for_each(ps, &output->state_cur->plane_list, link)
ps->complete = true;
drm_output_state_free(output->state_last);
output->state_last = NULL;
if (output->destroy_pending) {
output->destroy_pending = 0;
output->disable_pending = 0;
output->dpms_off_pending = 0;
drm_output_destroy(&output->base);
return;
} else if (output->disable_pending) {
output->disable_pending = 0;
output->dpms_off_pending = 0;
weston_output_disable(&output->base);
return;
} else if (output->dpms_off_pending) {
struct drm_pending_state *pending = drm_pending_state_alloc(b);
output->dpms_off_pending = 0;
drm_output_get_disable_state(pending, output);
drm_pending_state_apply_sync(pending);
return;
} else if (output->state_cur->dpms == WESTON_DPMS_OFF &&
output->base.repaint_status != REPAINT_AWAITING_COMPLETION) {
/* DPMS can happen to us either in the middle of a repaint
* cycle (when we have painted fresh content, only to throw it
* away for DPMS off), or at any other random point. If the
* latter is true, then we cannot go through finish_frame,
* because the repaint machinery does not expect this. */
return;
}
ts.tv_sec = sec;
ts.tv_nsec = usec * 1000;
weston_output_finish_frame(&output->base, &ts, flags);
/* We can't call this from frame_notify, because the output's
* repaint needed flag is cleared just after that */
if (output->recorder)
weston_output_schedule_repaint(&output->base);
}
/**
* Mark an output state as current on the output, i.e. it has been
* submitted to the kernel. The mode argument determines whether this
* update will be applied synchronously (e.g. when calling drmModeSetCrtc),
* or asynchronously (in which case we wait for events to complete).
*/
static void
drm_output_assign_state(struct drm_output_state *state,
enum drm_state_apply_mode mode)
{
struct drm_output *output = state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_plane_state *plane_state;
assert(!output->state_last);
if (mode == DRM_STATE_APPLY_ASYNC)
output->state_last = output->state_cur;
else
drm_output_state_free(output->state_cur);
wl_list_remove(&state->link);
wl_list_init(&state->link);
state->pending_state = NULL;
output->state_cur = state;
if (b->atomic_modeset && mode == DRM_STATE_APPLY_ASYNC)
output->atomic_complete_pending = 1;
/* Replace state_cur on each affected plane with the new state, being
* careful to dispose of orphaned (but only orphaned) previous state.
* If the previous state is not orphaned (still has an output_state
* attached), it will be disposed of by freeing the output_state. */
wl_list_for_each(plane_state, &state->plane_list, link) {
struct drm_plane *plane = plane_state->plane;
if (plane->state_cur && !plane->state_cur->output_state)
drm_plane_state_free(plane->state_cur, true);
plane->state_cur = plane_state;
if (mode != DRM_STATE_APPLY_ASYNC) {
plane_state->complete = true;
continue;
}
if (b->atomic_modeset)
continue;
if (plane->type == WDRM_PLANE_TYPE_OVERLAY)
output->vblank_pending++;
else if (plane->type == WDRM_PLANE_TYPE_PRIMARY)
output->page_flip_pending = 1;
}
}
enum drm_output_propose_state_mode {
DRM_OUTPUT_PROPOSE_STATE_MIXED, /**< mix renderer & planes */
DRM_OUTPUT_PROPOSE_STATE_RENDERER_ONLY, /**< only assign to renderer & cursor */
DRM_OUTPUT_PROPOSE_STATE_PLANES_ONLY, /**< no renderer use, only planes */
};
static struct drm_plane_state *
drm_output_prepare_scanout_view(struct drm_output_state *output_state,
struct weston_view *ev,
enum drm_output_propose_state_mode mode)
{
struct drm_output *output = output_state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_plane *scanout_plane = output->scanout_plane;
struct drm_plane_state *state;
struct drm_fb *fb;
pixman_box32_t *extents;
assert(!b->sprites_are_broken);
assert(mode == DRM_OUTPUT_PROPOSE_STATE_PLANES_ONLY);
/* Check the view spans exactly the output size, calculated in the
* logical co-ordinate space. */
extents = pixman_region32_extents(&ev->transform.boundingbox);
if (extents->x1 != output->base.x ||
extents->y1 != output->base.y ||
extents->x2 != output->base.x + output->base.width ||
extents->y2 != output->base.y + output->base.height)
return NULL;
if (ev->alpha != 1.0f)
return NULL;
fb = drm_fb_get_from_view(output_state, ev);
if (!fb)
return NULL;
/* Can't change formats with just a pageflip */
if (!b->atomic_modeset && fb->format->format != output->gbm_format) {
drm_fb_unref(fb);
return NULL;
}
state = drm_output_state_get_plane(output_state, scanout_plane);
/* The only way we can already have a buffer in the scanout plane is
* if we are in mixed mode, or if a client buffer has already been
* placed into scanout. The former case will never call into here,
* and in the latter case, the view must have been marked as occluded,
* meaning we should never have ended up here. */
assert(!state->fb);
state->fb = fb;
state->ev = ev;
state->output = output;
if (!drm_plane_state_coords_for_view(state, ev))
goto err;
if (state->dest_x != 0 || state->dest_y != 0 ||
state->dest_w != (unsigned) output->base.current_mode->width ||
state->dest_h != (unsigned) output->base.current_mode->height)
goto err;
/* The legacy API does not let us perform cropping or scaling. */
if (!b->atomic_modeset &&
(state->src_x != 0 || state->src_y != 0 ||
state->src_w != state->dest_w << 16 ||
state->src_h != state->dest_h << 16))
goto err;
/* In plane-only mode, we don't need to test the state now, as we
* will only test it once at the end. */
return state;
err:
drm_plane_state_put_back(state);
return NULL;
}
static struct drm_fb *
drm_output_render_gl(struct drm_output_state *state, pixman_region32_t *damage)
{
struct drm_output *output = state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct gbm_bo *bo;
struct drm_fb *ret;
output->base.compositor->renderer->repaint_output(&output->base,
damage);
bo = gbm_surface_lock_front_buffer(output->gbm_surface);
if (!bo) {
weston_log("failed to lock front buffer: %m\n");
return NULL;
}
/* The renderer always produces an opaque image. */
ret = drm_fb_get_from_bo(bo, b, true, BUFFER_GBM_SURFACE);
if (!ret) {
weston_log("failed to get drm_fb for bo\n");
gbm_surface_release_buffer(output->gbm_surface, bo);
return NULL;
}
ret->gbm_surface = output->gbm_surface;
return ret;
}
static struct drm_fb *
drm_output_render_pixman(struct drm_output_state *state,
pixman_region32_t *damage)
{
struct drm_output *output = state->output;
struct weston_compositor *ec = output->base.compositor;
output->current_image ^= 1;
pixman_renderer_output_set_buffer(&output->base,
output->image[output->current_image]);
pixman_renderer_output_set_hw_extra_damage(&output->base,
&output->previous_damage);
ec->renderer->repaint_output(&output->base, damage);
pixman_region32_copy(&output->previous_damage, damage);
return drm_fb_ref(output->dumb[output->current_image]);
}
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
static struct drm_fb *
drm_output_render_g2d(struct drm_output_state *state, pixman_region32_t *damage)
{
struct drm_output *output = state->output;
struct weston_compositor *ec = output->base.compositor;
pixman_region32_t total_damage, previous_damage;
pixman_region32_init(&total_damage);
pixman_region32_init(&previous_damage);
pixman_region32_copy(&previous_damage, damage);
pixman_region32_union(&total_damage, damage, &output->previous_damage);
pixman_region32_copy(&output->previous_damage, &previous_damage);
output->current_image ^= 1;
g2d_renderer->output_set_buffer(&output->base,
&output->g2d_image[output->current_image]);
ec->renderer->repaint_output(&output->base, &total_damage);
pixman_region32_fini(&total_damage);
pixman_region32_fini(&previous_damage);
return drm_fb_ref(output->dumb[output->current_image]);
}
#endif
#ifdef HAVE_GBM_MODIFIERS
static int
drm_get_gbm_alignment(struct drm_fb *fb)
{
int gbm_aligned = 1;
if (fb){
switch(fb->modifier) {
#if defined(ENABLE_IMXGPU)
case DRM_FORMAT_MOD_VIVANTE_SUPER_TILED_FC:
case DRM_FORMAT_MOD_VIVANTE_SUPER_TILED:
gbm_aligned = 64;
break;
#endif
default:
break;
}
}
return gbm_aligned;
}
#endif
static void
drm_output_render(struct drm_output_state *state, pixman_region32_t *damage)
{
struct drm_output *output = state->output;
struct weston_compositor *c = output->base.compositor;
struct drm_plane_state *scanout_state;
struct drm_plane *scanout_plane = output->scanout_plane;
struct drm_backend *b = to_drm_backend(c);
struct drm_fb *fb;
struct weston_config_section *section;
/* If we already have a client buffer promoted to scanout, then we don't
* want to render. */
scanout_state = drm_output_state_get_plane(state,
output->scanout_plane);
if (scanout_state->fb)
return;
#ifdef HAVE_GBM_MODIFIERS
int gbm_aligned = drm_get_gbm_alignment (scanout_plane->state_cur->fb);
#endif
if (!pixman_region32_not_empty(damage) &&
scanout_plane->state_cur->fb &&
(scanout_plane->state_cur->fb->type == BUFFER_GBM_SURFACE ||
scanout_plane->state_cur->fb->type == BUFFER_PIXMAN_DUMB) &&
#ifdef HAVE_GBM_MODIFIERS
scanout_plane->state_cur->fb->width ==
ALIGNTO(output->base.current_mode->width, gbm_aligned) &&
scanout_plane->state_cur->fb->height ==
ALIGNTO(output->base.current_mode->height, gbm_aligned)) {
#else
scanout_plane->state_cur->fb->width ==
output->base.current_mode->width &&
scanout_plane->state_cur->fb->height ==
output->base.current_mode->height) {
#endif
fb = drm_fb_ref(scanout_plane->state_cur->fb);
} else if (b->use_pixman) {
fb = drm_output_render_pixman(state, damage);
}
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_IMXG2D)
else if (b->use_g2d)
fb = drm_output_render_g2d(state, damage);
#endif
#if defined(ENABLE_OPENGL)
else
fb = drm_output_render_gl(state, damage);
#endif
#endif
if (!fb) {
drm_plane_state_put_back(scanout_state);
return;
}
scanout_state->fb = fb;
scanout_state->output = output;
scanout_state->src_x = 0;
scanout_state->src_y = 0;
scanout_state->src_w = output->base.current_mode->width << 16;
scanout_state->src_h = output->base.current_mode->height << 16;
scanout_state->dest_x = 0;
scanout_state->dest_y = 0;
scanout_state->dest_w = scanout_state->src_w >> 16;
scanout_state->dest_h = scanout_state->src_h >> 16;
section = weston_config_get_section(wet_get_config(c),
"shell", NULL, NULL);
if (section) {
char *size;
int n;
uint32_t width, height;
weston_config_section_get_string(section, "size", &size, NULL);
if(size){
n = sscanf(size, "%dx%d", &width, &height);
if (n == 2) {
if (scanout_state->src_w > (width << 16))
scanout_state->src_w = width << 16;
if (scanout_state->src_h > (height << 16))
scanout_state->src_h = height << 16;
}
}
}
pixman_region32_subtract(&c->primary_plane.damage,
&c->primary_plane.damage, damage);
}
static void
drm_output_set_gamma(struct weston_output *output_base,
uint16_t size, uint16_t *r, uint16_t *g, uint16_t *b)
{
int rc;
struct drm_output *output = to_drm_output(output_base);
struct drm_backend *backend =
to_drm_backend(output->base.compositor);
/* check */
if (output_base->gamma_size != size)
return;
rc = drmModeCrtcSetGamma(backend->drm.fd,
output->crtc_id,
size, r, g, b);
if (rc)
weston_log("set gamma failed: %m\n");
}
/* Determine the type of vblank synchronization to use for the output.
*
* The pipe parameter indicates which CRTC is in use. Knowing this, we
* can determine which vblank sequence type to use for it. Traditional
* cards had only two CRTCs, with CRTC 0 using no special flags, and
* CRTC 1 using DRM_VBLANK_SECONDARY. The first bit of the pipe
* parameter indicates this.
*
* Bits 1-5 of the pipe parameter are 5 bit wide pipe number between
* 0-31. If this is non-zero it indicates we're dealing with a
* multi-gpu situation and we need to calculate the vblank sync
* using DRM_BLANK_HIGH_CRTC_MASK.
*/
static unsigned int
drm_waitvblank_pipe(struct drm_output *output)
{
if (output->pipe > 1)
return (output->pipe << DRM_VBLANK_HIGH_CRTC_SHIFT) &
DRM_VBLANK_HIGH_CRTC_MASK;
else if (output->pipe > 0)
return DRM_VBLANK_SECONDARY;
else
return 0;
}
static int
drm_output_apply_state_legacy(struct drm_output_state *state)
{
struct drm_output *output = state->output;
struct drm_backend *backend = to_drm_backend(output->base.compositor);
struct drm_plane *scanout_plane = output->scanout_plane;
struct drm_property_info *dpms_prop;
struct drm_plane_state *scanout_state;
struct drm_plane_state *ps;
struct drm_mode *mode;
struct drm_head *head;
uint32_t connectors[MAX_CLONED_CONNECTORS];
int n_conn = 0;
struct timespec now;
int ret = 0;
wl_list_for_each(head, &output->base.head_list, base.output_link) {
assert(n_conn < MAX_CLONED_CONNECTORS);
connectors[n_conn++] = head->connector_id;
}
/* If disable_planes is set then assign_planes() wasn't
* called for this render, so we could still have a stale
* cursor plane set up.
*/
if (output->base.disable_planes) {
output->cursor_view = NULL;
if (output->cursor_plane) {
output->cursor_plane->base.x = INT32_MIN;
output->cursor_plane->base.y = INT32_MIN;
}
}
if (state->dpms != WESTON_DPMS_ON) {
wl_list_for_each(ps, &state->plane_list, link) {
struct drm_plane *p = ps->plane;
assert(ps->fb == NULL);
assert(ps->output == NULL);
if (p->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
ret = drmModeSetPlane(backend->drm.fd, p->plane_id,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0);
if (ret)
weston_log("drmModeSetPlane failed disable: %m\n");
}
if (output->cursor_plane) {
ret = drmModeSetCursor(backend->drm.fd, output->crtc_id,
0, 0, 0);
if (ret)
weston_log("drmModeSetCursor failed disable: %m\n");
}
ret = drmModeSetCrtc(backend->drm.fd, output->crtc_id, 0, 0, 0,
NULL, 0, NULL);
if (ret)
weston_log("drmModeSetCrtc failed disabling: %m\n");
drm_output_assign_state(state, DRM_STATE_APPLY_SYNC);
weston_compositor_read_presentation_clock(output->base.compositor, &now);
drm_output_update_complete(output,
WP_PRESENTATION_FEEDBACK_KIND_HW_COMPLETION,
now.tv_sec, now.tv_nsec / 1000);
return 0;
}
scanout_state =
drm_output_state_get_existing_plane(state, scanout_plane);
/* The legacy SetCrtc API doesn't allow us to do scaling, and the
* legacy PageFlip API doesn't allow us to do clipping either. */
assert(scanout_state->src_x == 0);
assert(scanout_state->src_y == 0);
assert(scanout_state->dest_x == 0);
assert(scanout_state->dest_y == 0);
mode = to_drm_mode(output->base.current_mode);
if (backend->state_invalid ||
!scanout_plane->state_cur->fb ||
scanout_plane->state_cur->fb->strides[0] !=
scanout_state->fb->strides[0]) {
ret = drmModeSetCrtc(backend->drm.fd, output->crtc_id,
scanout_state->fb->fb_id,
0, 0,
connectors, n_conn,
&mode->mode_info);
if (ret) {
weston_log("set mode failed: %m\n");
goto err;
}
}
if (drmModePageFlip(backend->drm.fd, output->crtc_id,
scanout_state->fb->fb_id,
DRM_MODE_PAGE_FLIP_EVENT, output) < 0) {
weston_log("queueing pageflip failed: %m\n");
goto err;
}
assert(!output->page_flip_pending);
if (output->pageflip_timer)
wl_event_source_timer_update(output->pageflip_timer,
backend->pageflip_timeout);
drm_output_set_cursor(state);
/*
* Now, update all the sprite surfaces
*/
wl_list_for_each(ps, &state->plane_list, link) {
uint32_t flags = 0, fb_id = 0;
drmVBlank vbl = {
.request.type = DRM_VBLANK_RELATIVE | DRM_VBLANK_EVENT,
.request.sequence = 1,
};
struct drm_plane *p = ps->plane;
if (p->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
assert(p->state_cur->complete);
assert(!!p->state_cur->output == !!p->state_cur->fb);
assert(!p->state_cur->output || p->state_cur->output == output);
assert(!ps->complete);
assert(!ps->output || ps->output == output);
assert(!!ps->output == !!ps->fb);
if (ps->fb && !backend->sprites_hidden)
fb_id = ps->fb->fb_id;
ret = drmModeSetPlane(backend->drm.fd, p->plane_id,
output->crtc_id, fb_id, flags,
ps->dest_x, ps->dest_y,
ps->dest_w, ps->dest_h,
ps->src_x, ps->src_y,
ps->src_w, ps->src_h);
if (ret)
weston_log("setplane failed: %d: %s\n",
ret, strerror(errno));
vbl.request.type |= drm_waitvblank_pipe(output);
/*
* Queue a vblank signal so we know when the surface
* becomes active on the display or has been replaced.
*/
vbl.request.signal = (unsigned long) ps;
ret = drmWaitVBlank(backend->drm.fd, &vbl);
if (ret) {
weston_log("vblank event request failed: %d: %s\n",
ret, strerror(errno));
}
}
if (state->dpms != output->state_cur->dpms) {
wl_list_for_each(head, &output->base.head_list, base.output_link) {
dpms_prop = &head->props_conn[WDRM_CONNECTOR_DPMS];
if (dpms_prop->prop_id == 0)
continue;
ret = drmModeConnectorSetProperty(backend->drm.fd,
head->connector_id,
dpms_prop->prop_id,
state->dpms);
if (ret) {
weston_log("DRM: DPMS: failed property set for %s\n",
head->base.name);
}
}
}
drm_output_assign_state(state, DRM_STATE_APPLY_ASYNC);
return 0;
err:
output->cursor_view = NULL;
drm_output_state_free(state);
return -1;
}
#ifdef HAVE_DRM_ATOMIC
static int
crtc_add_prop(drmModeAtomicReq *req, struct drm_output *output,
enum wdrm_crtc_property prop, uint64_t val)
{
struct drm_property_info *info = &output->props_crtc[prop];
int ret;
if (info->prop_id == 0)
return -1;
ret = drmModeAtomicAddProperty(req, output->crtc_id, info->prop_id,
val);
return (ret <= 0) ? -1 : 0;
}
static int
connector_add_prop(drmModeAtomicReq *req, struct drm_head *head,
enum wdrm_connector_property prop, uint64_t val)
{
struct drm_property_info *info = &head->props_conn[prop];
int ret;
if (info->prop_id == 0)
return -1;
ret = drmModeAtomicAddProperty(req, head->connector_id,
info->prop_id, val);
return (ret <= 0) ? -1 : 0;
}
static int
plane_add_prop(drmModeAtomicReq *req, struct drm_plane *plane,
enum wdrm_plane_property prop, uint64_t val)
{
struct drm_property_info *info = &plane->props[prop];
int ret;
if (info->prop_id == 0)
return -1;
ret = drmModeAtomicAddProperty(req, plane->plane_id, info->prop_id,
val);
return (ret <= 0) ? -1 : 0;
}
static int
drm_mode_ensure_blob(struct drm_backend *backend, struct drm_mode *mode)
{
int ret;
if (mode->blob_id)
return 0;
ret = drmModeCreatePropertyBlob(backend->drm.fd,
&mode->mode_info,
sizeof(mode->mode_info),
&mode->blob_id);
if (ret != 0)
weston_log("failed to create mode property blob: %m\n");
return ret;
}
static int
drm_output_apply_state_atomic(struct drm_output_state *state,
drmModeAtomicReq *req,
uint32_t *flags)
{
struct drm_output *output = state->output;
struct drm_backend *backend = to_drm_backend(output->base.compositor);
struct drm_plane_state *plane_state;
struct drm_mode *current_mode = to_drm_mode(output->base.current_mode);
struct drm_head *head;
int ret = 0;
if (state->dpms != output->state_cur->dpms)
*flags |= DRM_MODE_ATOMIC_ALLOW_MODESET;
if (state->dpms == WESTON_DPMS_ON) {
ret = drm_mode_ensure_blob(backend, current_mode);
if (ret != 0)
return ret;
ret |= crtc_add_prop(req, output, WDRM_CRTC_MODE_ID,
current_mode->blob_id);
ret |= crtc_add_prop(req, output, WDRM_CRTC_ACTIVE, 1);
/* No need for the DPMS property, since it is implicit in
* routing and CRTC activity. */
wl_list_for_each(head, &output->base.head_list, base.output_link) {
ret |= connector_add_prop(req, head, WDRM_CONNECTOR_CRTC_ID,
output->crtc_id);
}
if (backend->hdr_blob_id > 0) {
wl_list_for_each(head, &output->base.head_list, base.output_link) {
/* checking if the output driver this head */
if (head->base.output == &output->base)
connector_add_prop(req, head, WDRM_CONNECTOR_HDR10_METADATA,
backend->hdr_blob_id);
}
}
} else {
ret |= crtc_add_prop(req, output, WDRM_CRTC_MODE_ID, 0);
ret |= crtc_add_prop(req, output, WDRM_CRTC_ACTIVE, 0);
/* No need for the DPMS property, since it is implicit in
* routing and CRTC activity. */
wl_list_for_each(head, &output->base.head_list, base.output_link)
ret |= connector_add_prop(req, head, WDRM_CONNECTOR_CRTC_ID, 0);
}
if (ret != 0) {
weston_log("couldn't set atomic CRTC/connector state\n");
return ret;
}
wl_list_for_each(plane_state, &state->plane_list, link) {
struct drm_plane *plane = plane_state->plane;
ret |= plane_add_prop(req, plane, WDRM_PLANE_FB_ID,
plane_state->fb ? plane_state->fb->fb_id : 0);
ret |= plane_add_prop(req, plane, WDRM_PLANE_CRTC_ID,
plane_state->fb ? output->crtc_id : 0);
ret |= plane_add_prop(req, plane, WDRM_PLANE_SRC_X,
plane_state->src_x);
ret |= plane_add_prop(req, plane, WDRM_PLANE_SRC_Y,
plane_state->src_y);
ret |= plane_add_prop(req, plane, WDRM_PLANE_SRC_W,
plane_state->src_w);
ret |= plane_add_prop(req, plane, WDRM_PLANE_SRC_H,
plane_state->src_h);
ret |= plane_add_prop(req, plane, WDRM_PLANE_CRTC_X,
plane_state->dest_x);
ret |= plane_add_prop(req, plane, WDRM_PLANE_CRTC_Y,
plane_state->dest_y);
ret |= plane_add_prop(req, plane, WDRM_PLANE_CRTC_W,
plane_state->dest_w);
ret |= plane_add_prop(req, plane, WDRM_PLANE_CRTC_H,
plane_state->dest_h);
if (ret != 0) {
weston_log("couldn't set plane state\n");
return ret;
}
}
return 0;
}
/**
* Helper function used only by drm_pending_state_apply, with the same
* guarantees and constraints as that function.
*/
static int
drm_pending_state_apply_atomic(struct drm_pending_state *pending_state,
enum drm_state_apply_mode mode)
{
struct drm_backend *b = pending_state->backend;
struct drm_output_state *output_state, *tmp;
struct drm_plane *plane;
drmModeAtomicReq *req = drmModeAtomicAlloc();
uint32_t flags = 0;
int ret = 0;
drm_magic_t magic;
if (!req)
return -1;
if (b->state_invalid) {
struct weston_head *head_base;
struct drm_head *head;
uint32_t *unused;
int err;
/* If we need to reset all our state (e.g. because we've
* just started, or just been VT-switched in), explicitly
* disable all the CRTCs and connectors we aren't using. */
wl_list_for_each(head_base,
&b->compositor->head_list, compositor_link) {
struct drm_property_info *info;
if (weston_head_is_enabled(head_base))
continue;
head = to_drm_head(head_base);
info = &head->props_conn[WDRM_CONNECTOR_CRTC_ID];
err = drmModeAtomicAddProperty(req, head->connector_id,
info->prop_id, 0);
if (err <= 0)
ret = -1;
}
wl_array_for_each(unused, &b->unused_crtcs) {
struct drm_property_info infos[WDRM_CRTC__COUNT];
struct drm_property_info *info;
drmModeObjectProperties *props;
uint64_t active;
memset(infos, 0, sizeof(infos));
/* We can't emit a disable on a CRTC that's already
* off, as the kernel will refuse to generate an event
* for an off->off state and fail the commit.
*/
props = drmModeObjectGetProperties(b->drm.fd,
*unused,
DRM_MODE_OBJECT_CRTC);
if (!props) {
ret = -1;
continue;
}
drm_property_info_populate(b, crtc_props, infos,
WDRM_CRTC__COUNT,
props);
info = &infos[WDRM_CRTC_ACTIVE];
active = drm_property_get_value(info, props, 0);
drmModeFreeObjectProperties(props);
if (active == 0) {
drm_property_info_free(infos, WDRM_CRTC__COUNT);
continue;
}
err = drmModeAtomicAddProperty(req, *unused,
info->prop_id, 0);
if (err <= 0)
ret = -1;
info = &infos[WDRM_CRTC_MODE_ID];
err = drmModeAtomicAddProperty(req, *unused,
info->prop_id, 0);
if (err <= 0)
ret = -1;
drm_property_info_free(infos, WDRM_CRTC__COUNT);
}
/* Disable all the planes; planes which are being used will
* override this state in the output-state application. */
wl_list_for_each(plane, &b->plane_list, link) {
plane_add_prop(req, plane, WDRM_PLANE_CRTC_ID, 0);
plane_add_prop(req, plane, WDRM_PLANE_FB_ID, 0);
}
flags |= DRM_MODE_ATOMIC_ALLOW_MODESET;
}
wl_list_for_each(output_state, &pending_state->output_list, link) {
if (mode == DRM_STATE_APPLY_SYNC)
assert(output_state->dpms == WESTON_DPMS_OFF);
ret |= drm_output_apply_state_atomic(output_state, req, &flags);
}
if (ret != 0) {
weston_log("atomic: couldn't compile atomic state\n");
goto out;
}
switch (mode) {
case DRM_STATE_APPLY_SYNC:
break;
case DRM_STATE_APPLY_ASYNC:
flags |= DRM_MODE_PAGE_FLIP_EVENT | DRM_MODE_ATOMIC_NONBLOCK;
break;
case DRM_STATE_TEST_ONLY:
flags |= DRM_MODE_ATOMIC_TEST_ONLY;
break;
}
/*drm master was set by systemd in PM test, try to set the master back.*/
if (!(drmGetMagic(b->drm.fd, &magic) == 0 &&
drmAuthMagic(b->drm.fd, magic) == 0)) {
drmSetMaster(b->drm.fd);
}
ret = drmModeAtomicCommit(b->drm.fd, req, flags, b);
/* Test commits do not take ownership of the state; return
* without freeing here. */
if (mode == DRM_STATE_TEST_ONLY) {
drmModeAtomicFree(req);
return ret;
}
if (ret != 0) {
weston_log("atomic: couldn't commit new state: %m\n");
goto out;
}
wl_list_for_each_safe(output_state, tmp, &pending_state->output_list,
link)
drm_output_assign_state(output_state, mode);
b->state_invalid = false;
assert(wl_list_empty(&pending_state->output_list));
out:
if (b->hdr_blob_id > 0) {
drmModeDestroyPropertyBlob (b->drm.fd, b->hdr_blob_id);
b->hdr_blob_id = 0;
}
drmModeAtomicFree(req);
drm_pending_state_free(pending_state);
return ret;
}
#endif
/**
* Tests a pending state, to see if the kernel will accept the update as
* constructed.
*
* Using atomic modesetting, the kernel performs the same checks as it would
* on a real commit, returning success or failure without actually modifying
* the running state. It does not return -EBUSY if there are pending updates
* in flight, so states may be tested at any point, however this means a
* state which passed testing may fail on a real commit if the timing is not
* respected (e.g. committing before the previous commit has completed).
*
* Without atomic modesetting, we have no way to check, so we optimistically
* claim it will work.
*
* Unlike drm_pending_state_apply() and drm_pending_state_apply_sync(), this
* function does _not_ take ownership of pending_state, nor does it clear
* state_invalid.
*/
static int
drm_pending_state_test(struct drm_pending_state *pending_state)
{
#ifdef HAVE_DRM_ATOMIC
struct drm_backend *b = pending_state->backend;
if (b->atomic_modeset)
return drm_pending_state_apply_atomic(pending_state,
DRM_STATE_TEST_ONLY);
#endif
/* We have no way to test state before application on the legacy
* modesetting API, so just claim it succeeded. */
return 0;
}
/**
* Applies all of a pending_state asynchronously: the primary entry point for
* applying KMS state to a device. Updates the state for all outputs in the
* pending_state, as well as disabling any unclaimed outputs.
*
* Unconditionally takes ownership of pending_state, and clears state_invalid.
*/
static int
drm_pending_state_apply(struct drm_pending_state *pending_state)
{
struct drm_backend *b = pending_state->backend;
struct drm_output_state *output_state, *tmp;
uint32_t *unused;
#ifdef HAVE_DRM_ATOMIC
if (b->atomic_modeset)
return drm_pending_state_apply_atomic(pending_state,
DRM_STATE_APPLY_ASYNC);
#endif
if (b->state_invalid) {
/* If we need to reset all our state (e.g. because we've
* just started, or just been VT-switched in), explicitly
* disable all the CRTCs we aren't using. This also disables
* all connectors on these CRTCs, so we don't need to do that
* separately with the pre-atomic API. */
wl_array_for_each(unused, &b->unused_crtcs)
drmModeSetCrtc(b->drm.fd, *unused, 0, 0, 0, NULL, 0,
NULL);
}
wl_list_for_each_safe(output_state, tmp, &pending_state->output_list,
link) {
struct drm_output *output = output_state->output;
int ret;
ret = drm_output_apply_state_legacy(output_state);
if (ret != 0) {
weston_log("Couldn't apply state for output %s\n",
output->base.name);
}
}
b->state_invalid = false;
assert(wl_list_empty(&pending_state->output_list));
drm_pending_state_free(pending_state);
return 0;
}
/**
* The synchronous version of drm_pending_state_apply. May only be used to
* disable outputs. Does so synchronously: the request is guaranteed to have
* completed on return, and the output will not be touched afterwards.
*
* Unconditionally takes ownership of pending_state, and clears state_invalid.
*/
static int
drm_pending_state_apply_sync(struct drm_pending_state *pending_state)
{
struct drm_backend *b = pending_state->backend;
struct drm_output_state *output_state, *tmp;
uint32_t *unused;
#ifdef HAVE_DRM_ATOMIC
if (b->atomic_modeset)
return drm_pending_state_apply_atomic(pending_state,
DRM_STATE_APPLY_SYNC);
#endif
if (b->state_invalid) {
/* If we need to reset all our state (e.g. because we've
* just started, or just been VT-switched in), explicitly
* disable all the CRTCs we aren't using. This also disables
* all connectors on these CRTCs, so we don't need to do that
* separately with the pre-atomic API. */
wl_array_for_each(unused, &b->unused_crtcs)
drmModeSetCrtc(b->drm.fd, *unused, 0, 0, 0, NULL, 0,
NULL);
}
wl_list_for_each_safe(output_state, tmp, &pending_state->output_list,
link) {
int ret;
assert(output_state->dpms == WESTON_DPMS_OFF);
ret = drm_output_apply_state_legacy(output_state);
if (ret != 0) {
weston_log("Couldn't apply state for output %s\n",
output_state->output->base.name);
}
}
b->state_invalid = false;
assert(wl_list_empty(&pending_state->output_list));
drm_pending_state_free(pending_state);
return 0;
}
static int
drm_output_repaint(struct weston_output *output_base,
pixman_region32_t *damage,
void *repaint_data)
{
struct drm_pending_state *pending_state = repaint_data;
struct drm_output *output = to_drm_output(output_base);
struct drm_output_state *state = NULL;
struct drm_plane_state *scanout_state;
if (output->disable_pending || output->destroy_pending)
goto err;
assert(!output->state_last);
/* If planes have been disabled in the core, we might not have
* hit assign_planes at all, so might not have valid output state
* here. */
state = drm_pending_state_get_output(pending_state, output);
if (!state)
state = drm_output_state_duplicate(output->state_cur,
pending_state,
DRM_OUTPUT_STATE_CLEAR_PLANES);
state->dpms = WESTON_DPMS_ON;
drm_output_render(state, damage);
scanout_state = drm_output_state_get_plane(state,
output->scanout_plane);
if (!scanout_state || !scanout_state->fb)
goto err;
return 0;
err:
drm_output_state_free(state);
return -1;
}
static void
drm_output_start_repaint_loop(struct weston_output *output_base)
{
struct drm_output *output = to_drm_output(output_base);
struct drm_pending_state *pending_state;
struct drm_plane *scanout_plane = output->scanout_plane;
struct drm_backend *backend =
to_drm_backend(output_base->compositor);
struct timespec ts, tnow;
struct timespec vbl2now;
int64_t refresh_nsec;
int ret;
drmVBlank vbl = {
.request.type = DRM_VBLANK_RELATIVE,
.request.sequence = 0,
.request.signal = 0,
};
if (output->disable_pending || output->destroy_pending)
return;
if (!output->scanout_plane->state_cur->fb) {
/* We can't page flip if there's no mode set */
goto finish_frame;
}
/* Need to smash all state in from scratch; current timings might not
* be what we want, page flip might not work, etc.
*/
if (backend->state_invalid)
goto finish_frame;
assert(scanout_plane->state_cur->output == output);
/* Try to get current msc and timestamp via instant query */
vbl.request.type |= drm_waitvblank_pipe(output);
ret = drmWaitVBlank(backend->drm.fd, &vbl);
/* Error ret or zero timestamp means failure to get valid timestamp */
if ((ret == 0) && (vbl.reply.tval_sec > 0 || vbl.reply.tval_usec > 0)) {
ts.tv_sec = vbl.reply.tval_sec;
ts.tv_nsec = vbl.reply.tval_usec * 1000;
/* Valid timestamp for most recent vblank - not stale?
* Stale ts could happen on Linux 3.17+, so make sure it
* is not older than 1 refresh duration since now.
*/
weston_compositor_read_presentation_clock(backend->compositor,
&tnow);
timespec_sub(&vbl2now, &tnow, &ts);
refresh_nsec =
millihz_to_nsec(output->base.current_mode->refresh);
if (timespec_to_nsec(&vbl2now) < refresh_nsec) {
drm_output_update_msc(output, vbl.reply.sequence);
weston_output_finish_frame(output_base, &ts,
WP_PRESENTATION_FEEDBACK_INVALID);
return;
}
}
/* Immediate query didn't provide valid timestamp.
* Use pageflip fallback.
*/
assert(!output->page_flip_pending);
assert(!output->state_last);
pending_state = drm_pending_state_alloc(backend);
drm_output_state_duplicate(output->state_cur, pending_state,
DRM_OUTPUT_STATE_PRESERVE_PLANES);
ret = drm_pending_state_apply(pending_state);
if (ret != 0) {
weston_log("applying repaint-start state failed: %m\n");
goto finish_frame;
}
return;
finish_frame:
/* if we cannot page-flip, immediately finish frame */
weston_output_finish_frame(output_base, NULL,
WP_PRESENTATION_FEEDBACK_INVALID);
}
static void
drm_output_update_msc(struct drm_output *output, unsigned int seq)
{
uint64_t msc_hi = output->base.msc >> 32;
if (seq < (output->base.msc & 0xffffffff))
msc_hi++;
output->base.msc = (msc_hi << 32) + seq;
}
static void
vblank_handler(int fd, unsigned int frame, unsigned int sec, unsigned int usec,
void *data)
{
struct drm_plane_state *ps = (struct drm_plane_state *) data;
struct drm_output_state *os = ps->output_state;
struct drm_output *output = os->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
uint32_t flags = WP_PRESENTATION_FEEDBACK_KIND_HW_COMPLETION |
WP_PRESENTATION_FEEDBACK_KIND_HW_CLOCK;
assert(!b->atomic_modeset);
drm_output_update_msc(output, frame);
output->vblank_pending--;
assert(output->vblank_pending >= 0);
assert(ps->fb);
if (output->page_flip_pending || output->vblank_pending)
return;
drm_output_update_complete(output, flags, sec, usec);
}
static void
page_flip_handler(int fd, unsigned int frame,
unsigned int sec, unsigned int usec, void *data)
{
struct drm_output *output = data;
struct drm_backend *b = to_drm_backend(output->base.compositor);
uint32_t flags = WP_PRESENTATION_FEEDBACK_KIND_VSYNC |
WP_PRESENTATION_FEEDBACK_KIND_HW_COMPLETION |
WP_PRESENTATION_FEEDBACK_KIND_HW_CLOCK;
drm_output_update_msc(output, frame);
assert(!b->atomic_modeset);
assert(output->page_flip_pending);
output->page_flip_pending = 0;
if (output->vblank_pending)
return;
drm_output_update_complete(output, flags, sec, usec);
}
/**
* Begin a new repaint cycle
*
* Called by the core compositor at the beginning of a repaint cycle. Creates
* a new pending_state structure to own any output state created by individual
* output repaint functions until the repaint is flushed or cancelled.
*/
static void *
drm_repaint_begin(struct weston_compositor *compositor)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_pending_state *ret;
ret = drm_pending_state_alloc(b);
b->repaint_data = ret;
return ret;
}
/**
* Flush a repaint set
*
* Called by the core compositor when a repaint cycle has been completed
* and should be flushed. Frees the pending state, transitioning ownership
* of the output state from the pending state, to the update itself. When
* the update completes (see drm_output_update_complete), the output
* state will be freed.
*/
static void
drm_repaint_flush(struct weston_compositor *compositor, void *repaint_data)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_pending_state *pending_state = repaint_data;
drm_pending_state_apply(pending_state);
b->repaint_data = NULL;
}
/**
* Cancel a repaint set
*
* Called by the core compositor when a repaint has finished, so the data
* held across the repaint cycle should be discarded.
*/
static void
drm_repaint_cancel(struct weston_compositor *compositor, void *repaint_data)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_pending_state *pending_state = repaint_data;
drm_pending_state_free(pending_state);
b->repaint_data = NULL;
}
static void
drm_query_dmabuf_formats(struct weston_compositor *compositor, int **formats, int *num_formats)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_plane *p;
uint64_t has_prime;
uint32_t i, j;
int ret;
assert(formats);
assert(num_formats);
*num_formats = 0;
ret = drmGetCap (b->drm.fd, DRM_CAP_PRIME, &has_prime);
if (ret || !(bool) (has_prime & DRM_PRIME_CAP_IMPORT)) {
weston_log("drm backend not support import DMABUF\n");
return;
}
wl_list_for_each(p, &b->plane_list, link) {
if (p->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
*formats = calloc(p->count_formats, sizeof(int));
if (*formats == NULL) {
*num_formats = 0;
continue;
}
j = 0;
for (i = 0; i < p->count_formats; i++)
(*formats)[j++] = p->formats[i].format;
*num_formats = j;
if (*num_formats > 0)
break;
}
}
/**
* Test if drm driver can import dmabuf
*
* called by compositor when a dmabuf comes to test if this buffer
* can used by drm driver directly
*/
static bool
drm_import_dmabuf(struct weston_compositor *compositor,
struct linux_dmabuf_buffer *dmabuf)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_plane *p;
uint64_t has_prime;
uint32_t i;
int ret;
ret = drmGetCap (b->drm.fd, DRM_CAP_PRIME, &has_prime);
if (ret || !(bool) (has_prime & DRM_PRIME_CAP_IMPORT)) {
weston_log("drm backend not support import DMABUF\n");
return false;
}
wl_list_for_each(p, &b->plane_list, link) {
if (p->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
for (i = 0; i < p->count_formats; i++) {
if (p->formats[i].format == dmabuf->attributes.format
&& dmabuf->attributes.format == DRM_FORMAT_P010)
return true;
}
}
return false;
}
#ifdef HAVE_DRM_ATOMIC
static void
atomic_flip_handler(int fd, unsigned int frame, unsigned int sec,
unsigned int usec, unsigned int crtc_id, void *data)
{
struct drm_backend *b = data;
struct drm_output *output = drm_output_find_by_crtc(b, crtc_id);
uint32_t flags = WP_PRESENTATION_FEEDBACK_KIND_VSYNC |
WP_PRESENTATION_FEEDBACK_KIND_HW_COMPLETION |
WP_PRESENTATION_FEEDBACK_KIND_HW_CLOCK;
/* During the initial modeset, we can disable CRTCs which we don't
* actually handle during normal operation; this will give us events
* for unknown outputs. Ignore them. */
if (!output || !output->base.enabled)
return;
drm_output_update_msc(output, frame);
assert(b->atomic_modeset);
assert(output->atomic_complete_pending);
output->atomic_complete_pending = 0;
drm_output_update_complete(output, flags, sec, usec);
}
#endif
static struct drm_plane_state *
drm_output_prepare_overlay_view(struct drm_output_state *output_state,
struct weston_view *ev,
enum drm_output_propose_state_mode mode)
{
struct drm_output *output = output_state->output;
struct weston_compositor *ec = output->base.compositor;
struct drm_backend *b = to_drm_backend(ec);
struct drm_plane *p;
struct drm_plane_state *state = NULL;
struct drm_fb *fb;
unsigned int i;
int ret;
assert(!b->sprites_are_broken);
fb = drm_fb_get_from_view(output_state, ev);
if (!fb)
return NULL;
wl_list_for_each(p, &b->plane_list, link) {
if (p->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
if (!drm_plane_is_available(p, output))
continue;
/* Check whether the format is supported */
for (i = 0; i < p->count_formats; i++) {
unsigned int j;
if (p->formats[i].format != fb->format->format)
continue;
if (fb->modifier == DRM_FORMAT_MOD_INVALID)
break;
for (j = 0; j < p->formats[i].count_modifiers; j++) {
if (p->formats[i].modifiers[j] == fb->modifier)
break;
}
if (j != p->formats[i].count_modifiers)
break;
}
if (i == p->count_formats)
continue;
state = drm_output_state_get_plane(output_state, p);
if (state->fb) {
state = NULL;
continue;
}
state->ev = ev;
state->output = output;
/* We hold one reference for the lifetime of this function;
* from calling drm_fb_get_from_view, to the out label where
* we unconditionally drop the reference. So, we take another
* reference here to live within the state. */
state->fb = drm_fb_ref(fb);
if (!drm_plane_state_coords_for_view(state, ev)) {
drm_fb_unref(fb);
drm_plane_state_put_back(state);
state = NULL;
continue;
}
if (!b->atomic_modeset &&
(state->src_w != state->dest_w << 16 ||
state->src_h != state->dest_h << 16)) {
drm_fb_unref(fb);
drm_plane_state_put_back(state);
state = NULL;
continue;
}
/* In planes-only mode, we don't have an incremental state to
* test against, so we just hope it'll work. */
if (mode == DRM_OUTPUT_PROPOSE_STATE_PLANES_ONLY)
goto out;
ret = drm_pending_state_test(output_state->pending_state);
if (ret == 0)
goto out;
drm_plane_state_put_back(state);
state = NULL;
}
out:
drm_fb_unref(fb);
return state;
}
/**
* Update the image for the current cursor surface
*
* @param plane_state DRM cursor plane state
* @param ev Source view for cursor
*/
static void
cursor_bo_update(struct drm_plane_state *plane_state, struct weston_view *ev)
{
struct drm_backend *b = plane_state->plane->backend;
struct gbm_bo *bo = plane_state->fb->bo;
struct weston_buffer *buffer = ev->surface->buffer_ref.buffer;
uint32_t buf[b->cursor_width * b->cursor_height];
int32_t stride;
uint8_t *s;
int i;
assert(buffer && buffer->shm_buffer);
assert(buffer->shm_buffer == wl_shm_buffer_get(buffer->resource));
assert(buffer->width <= b->cursor_width);
assert(buffer->height <= b->cursor_height);
memset(buf, 0, sizeof buf);
stride = wl_shm_buffer_get_stride(buffer->shm_buffer);
s = wl_shm_buffer_get_data(buffer->shm_buffer);
wl_shm_buffer_begin_access(buffer->shm_buffer);
for (i = 0; i < buffer->height; i++)
memcpy(buf + i * b->cursor_width,
s + i * stride,
buffer->width * 4);
wl_shm_buffer_end_access(buffer->shm_buffer);
if (gbm_bo_write(bo, buf, sizeof buf) < 0)
weston_log("failed update cursor: %m\n");
}
static struct drm_plane_state *
drm_output_prepare_cursor_view(struct drm_output_state *output_state,
struct weston_view *ev)
{
struct drm_output *output = output_state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_plane *plane = output->cursor_plane;
struct drm_plane_state *plane_state;
struct wl_shm_buffer *shmbuf;
bool needs_update = false;
assert(!b->cursors_are_broken);
if (!plane)
return NULL;
if (!plane->state_cur->complete)
return NULL;
if (plane->state_cur->output && plane->state_cur->output != output)
return NULL;
/* We use GBM to import SHM buffers. */
if (b->gbm == NULL)
return NULL;
if (ev->surface->buffer_ref.buffer == NULL)
return NULL;
shmbuf = wl_shm_buffer_get(ev->surface->buffer_ref.buffer->resource);
if (!shmbuf)
return NULL;
if (wl_shm_buffer_get_format(shmbuf) != WL_SHM_FORMAT_ARGB8888)
return NULL;
plane_state =
drm_output_state_get_plane(output_state, output->cursor_plane);
if (plane_state && plane_state->fb)
return NULL;
/* We can't scale with the legacy API, and we don't try to account for
* simple cropping/translation in cursor_bo_update. */
plane_state->output = output;
if (!drm_plane_state_coords_for_view(plane_state, ev))
goto err;
if (plane_state->src_x != 0 || plane_state->src_y != 0 ||
plane_state->src_w > (unsigned) b->cursor_width << 16 ||
plane_state->src_h > (unsigned) b->cursor_height << 16 ||
plane_state->src_w != plane_state->dest_w << 16 ||
plane_state->src_h != plane_state->dest_h << 16)
goto err;
/* Since we're setting plane state up front, we need to work out
* whether or not we need to upload a new cursor. We can't use the
* plane damage, since the planes haven't actually been calculated
* yet: instead try to figure it out directly. KMS cursor planes are
* pretty unique here, in that they lie partway between a Weston plane
* (direct scanout) and a renderer. */
if (ev != output->cursor_view ||
pixman_region32_not_empty(&ev->surface->damage)) {
output->current_cursor++;
output->current_cursor =
output->current_cursor %
ARRAY_LENGTH(output->gbm_cursor_fb);
needs_update = true;
}
output->cursor_view = ev;
plane_state->ev = ev;
plane_state->fb =
drm_fb_ref(output->gbm_cursor_fb[output->current_cursor]);
if (needs_update)
cursor_bo_update(plane_state, ev);
/* The cursor API is somewhat special: in cursor_bo_update(), we upload
* a buffer which is always cursor_width x cursor_height, even if the
* surface we want to promote is actually smaller than this. Manually
* mangle the plane state to deal with this. */
plane_state->src_w = b->cursor_width << 16;
plane_state->src_h = b->cursor_height << 16;
plane_state->dest_w = b->cursor_width;
plane_state->dest_h = b->cursor_height;
return plane_state;
err:
drm_plane_state_put_back(plane_state);
return NULL;
}
static void
drm_output_set_cursor(struct drm_output_state *output_state)
{
struct drm_output *output = output_state->output;
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_plane *plane = output->cursor_plane;
struct drm_plane_state *state;
EGLint handle;
struct gbm_bo *bo;
if (!plane)
return;
state = drm_output_state_get_existing_plane(output_state, plane);
if (!state)
return;
if (!state->fb) {
pixman_region32_fini(&plane->base.damage);
pixman_region32_init(&plane->base.damage);
drmModeSetCursor(b->drm.fd, output->crtc_id, 0, 0, 0);
return;
}
assert(state->fb == output->gbm_cursor_fb[output->current_cursor]);
assert(!plane->state_cur->output || plane->state_cur->output == output);
if (plane->state_cur->fb != state->fb) {
bo = state->fb->bo;
handle = gbm_bo_get_handle(bo).s32;
if (drmModeSetCursor(b->drm.fd, output->crtc_id, handle,
b->cursor_width, b->cursor_height)) {
weston_log("failed to set cursor: %m\n");
goto err;
}
}
pixman_region32_fini(&plane->base.damage);
pixman_region32_init(&plane->base.damage);
if (drmModeMoveCursor(b->drm.fd, output->crtc_id,
state->dest_x, state->dest_y)) {
weston_log("failed to move cursor: %m\n");
goto err;
}
return;
err:
b->cursors_are_broken = 1;
drmModeSetCursor(b->drm.fd, output->crtc_id, 0, 0, 0);
}
static struct drm_output_state *
drm_output_propose_state(struct weston_output *output_base,
struct drm_pending_state *pending_state,
enum drm_output_propose_state_mode mode)
{
struct drm_output *output = to_drm_output(output_base);
struct drm_backend *b = to_drm_backend(output->base.compositor);
struct drm_output_state *state;
struct drm_plane_state *scanout_state = NULL;
struct weston_view *ev;
pixman_region32_t surface_overlap, renderer_region, occluded_region;
bool planes_ok = (mode != DRM_OUTPUT_PROPOSE_STATE_RENDERER_ONLY);
bool renderer_ok = (mode != DRM_OUTPUT_PROPOSE_STATE_PLANES_ONLY);
int ret;
assert(!output->state_last);
state = drm_output_state_duplicate(output->state_cur,
pending_state,
DRM_OUTPUT_STATE_CLEAR_PLANES);
/* We implement mixed mode by progressively creating and testing
* incremental states, of scanout + overlay + cursor. Since we
* walk our views top to bottom, the scanout plane is last, however
* we always need it in our scene for the test modeset to be
* meaningful. To do this, we steal a reference to the last
* renderer framebuffer we have, if we think it's basically
* compatible. If we don't have that, then we conservatively fall
* back to only using the renderer for this repaint. */
if (mode == DRM_OUTPUT_PROPOSE_STATE_MIXED) {
struct drm_plane *plane = output->scanout_plane;
struct drm_fb *scanout_fb = plane->state_cur->fb;
if (!scanout_fb ||
(scanout_fb->type != BUFFER_GBM_SURFACE &&
scanout_fb->type != BUFFER_PIXMAN_DUMB)) {
drm_output_state_free(state);
return NULL;
}
#ifdef HAVE_GBM_MODIFIERS
int gbm_aligned = drm_get_gbm_alignment (scanout_fb);
if (scanout_fb->width != ALIGNTO(output_base->current_mode->width, gbm_aligned) ||
scanout_fb->height != ALIGNTO(output_base->current_mode->height, gbm_aligned)) {
#else
if (scanout_fb->width != output_base->current_mode->width ||
scanout_fb->height != output_base->current_mode->height) {
#endif
drm_output_state_free(state);
return NULL;
}
scanout_state = drm_plane_state_duplicate(state,
plane->state_cur);
}
/*
* Find a surface for each sprite in the output using some heuristics:
* 1) size
* 2) frequency of update
* 3) opacity (though some hw might support alpha blending)
* 4) clipping (this can be fixed with color keys)
*
* The idea is to save on blitting since this should save power.
* If we can get a large video surface on the sprite for example,
* the main display surface may not need to update at all, and
* the client buffer can be used directly for the sprite surface
* as we do for flipping full screen surfaces.
*/
pixman_region32_init(&renderer_region);
pixman_region32_init(&occluded_region);
wl_list_for_each(ev, &output_base->compositor->view_list, link) {
struct drm_plane_state *ps = NULL;
bool force_renderer = false;
pixman_region32_t clipped_view;
bool totally_occluded = false;
bool overlay_occluded = false;
struct linux_dmabuf_buffer *dmabuf = NULL;
/* If this view doesn't touch our output at all, there's no
* reason to do anything with it. */
if (!(ev->output_mask & (1u << output->base.id)))
continue;
/* We only assign planes to views which are exclusively present
* on our output. */
if (ev->output_mask != (1u << output->base.id))
force_renderer = true;
if (!ev->surface->buffer_ref.buffer)
force_renderer = true;
else {
struct weston_buffer *buffer = ev->surface->buffer_ref.buffer;
dmabuf = linux_dmabuf_buffer_get(buffer->resource);
if (dmabuf) {
if (dmabuf->attributes.format == DRM_FORMAT_NV12
|| dmabuf->attributes.format == DRM_FORMAT_P010
|| dmabuf->attributes.format == DRM_FORMAT_YUYV) {
ps = drm_output_prepare_overlay_view(state, ev, mode);
if (ps)
goto next_view;
else
force_renderer = true;
} else
force_renderer = true;
}
}
/* Ignore views we know to be totally occluded. */
pixman_region32_init(&clipped_view);
pixman_region32_intersect(&clipped_view,
&ev->transform.boundingbox,
&output->base.region);
pixman_region32_init(&surface_overlap);
pixman_region32_subtract(&surface_overlap, &clipped_view,
&occluded_region);
totally_occluded = !pixman_region32_not_empty(&surface_overlap);
if (totally_occluded) {
pixman_region32_fini(&surface_overlap);
pixman_region32_fini(&clipped_view);
continue;
}
/* Since we process views from top to bottom, we know that if
* the view intersects the calculated renderer region, it must
* be part of, or occluded by, it, and cannot go on a plane. */
pixman_region32_intersect(&surface_overlap, &renderer_region,
&clipped_view);
if (pixman_region32_not_empty(&surface_overlap))
force_renderer = true;
/* We do not control the stacking order of overlay planes;
* the scanout plane is strictly stacked bottom and the cursor
* plane top, but the ordering of overlay planes with respect
* to each other is undefined. Make sure we do not have two
* planes overlapping each other. */
pixman_region32_intersect(&surface_overlap, &occluded_region,
&clipped_view);
if (pixman_region32_not_empty(&surface_overlap))
overlay_occluded = true;
pixman_region32_fini(&surface_overlap);
/* The cursor plane is 'special' in the sense that we can still
* place it in the legacy API, and we gate that with a separate
* cursors_are_broken flag. */
if (!force_renderer && !overlay_occluded && !b->cursors_are_broken)
ps = drm_output_prepare_cursor_view(state, ev);
/* If sprites are disabled or the view is not fully opaque, we
* must put the view into the renderer - unless it has already
* been placed in the cursor plane, which can handle alpha. */
if (!ps && !planes_ok)
force_renderer = true;
if (!ps && !drm_view_is_opaque(ev))
force_renderer = true;
/* Only try to place scanout surfaces in planes-only mode; in
* mixed mode, we have already failed to place a view on the
* scanout surface, forcing usage of the renderer on the
* scanout plane. */
if (!ps && !force_renderer && !renderer_ok)
ps = drm_output_prepare_scanout_view(state, ev, mode);
if (!ps && !overlay_occluded && !force_renderer)
ps = drm_output_prepare_overlay_view(state, ev, mode);
next_view:
if (ps) {
/* If we have been assigned to an overlay or scanout
* plane, add this area to the occluded region, so
* other views are known to be behind it. The cursor
* plane, however, is special, in that it blends with
* the content underneath it: the area should neither
* be added to the renderer region nor the occluded
* region. */
if (ps->plane->type != WDRM_PLANE_TYPE_CURSOR) {
pixman_region32_union(&occluded_region,
&occluded_region,
&clipped_view);
pixman_region32_fini(&clipped_view);
}
continue;
}
/* We have been assigned to the primary (renderer) plane:
* check if this is OK, and add ourselves to the renderer
* region if so. */
if (!renderer_ok) {
pixman_region32_fini(&clipped_view);
goto err_region;
}
pixman_region32_union(&renderer_region,
&renderer_region,
&clipped_view);
pixman_region32_fini(&clipped_view);
}
pixman_region32_fini(&renderer_region);
pixman_region32_fini(&occluded_region);
/* In renderer-only mode, we can't test the state as we don't have a
* renderer buffer yet. */
if (mode == DRM_OUTPUT_PROPOSE_STATE_RENDERER_ONLY)
return state;
/* Check to see if this state will actually work. */
ret = drm_pending_state_test(state->pending_state);
if (ret != 0)
goto err;
/* Counterpart to duplicating scanout state at the top of this
* function: if we have taken a renderer framebuffer and placed it in
* the pending state in order to incrementally test overlay planes,
* remove it now. */
if (mode == DRM_OUTPUT_PROPOSE_STATE_MIXED) {
assert(scanout_state->fb->type == BUFFER_GBM_SURFACE ||
scanout_state->fb->type == BUFFER_PIXMAN_DUMB);
drm_plane_state_put_back(scanout_state);
}
return state;
err_region:
pixman_region32_fini(&renderer_region);
pixman_region32_fini(&occluded_region);
err:
drm_output_state_free(state);
return NULL;
}
static void
drm_assign_planes(struct weston_output *output_base, void *repaint_data)
{
struct drm_backend *b = to_drm_backend(output_base->compositor);
struct drm_pending_state *pending_state = repaint_data;
struct drm_output *output = to_drm_output(output_base);
struct drm_output_state *state = NULL;
struct drm_plane_state *plane_state;
struct weston_view *ev;
struct weston_plane *primary = &output_base->compositor->primary_plane;
if (!b->sprites_are_broken) {
state = drm_output_propose_state(output_base, pending_state,
DRM_OUTPUT_PROPOSE_STATE_PLANES_ONLY);
if (!state)
state = drm_output_propose_state(output_base, pending_state,
DRM_OUTPUT_PROPOSE_STATE_MIXED);
}
if (!state)
state = drm_output_propose_state(output_base, pending_state,
DRM_OUTPUT_PROPOSE_STATE_RENDERER_ONLY);
assert(state);
wl_list_for_each(ev, &output_base->compositor->view_list, link) {
struct drm_plane *target_plane = NULL;
/* If this view doesn't touch our output at all, there's no
* reason to do anything with it. */
if (!(ev->output_mask & (1u << output->base.id)))
continue;
/* Test whether this buffer can ever go into a plane:
* non-shm, or small enough to be a cursor.
*
* Also, keep a reference when using the pixman renderer.
* That makes it possible to do a seamless switch to the GL
* renderer and since the pixman renderer keeps a reference
* to the buffer anyway, there is no side effects.
*/
if (b->use_pixman ||
(ev->surface->buffer_ref.buffer &&
(!wl_shm_buffer_get(ev->surface->buffer_ref.buffer->resource) ||
(ev->surface->width <= b->cursor_width &&
ev->surface->height <= b->cursor_height))))
ev->surface->keep_buffer = true;
else
ev->surface->keep_buffer = false;
/* This is a bit unpleasant, but lacking a temporary place to
* hang a plane off the view, we have to do a nested walk.
* Our first-order iteration has to be planes rather than
* views, because otherwise we won't reset views which were
* previously on planes to being on the primary plane. */
wl_list_for_each(plane_state, &state->plane_list, link) {
if (plane_state->ev == ev) {
plane_state->ev = NULL;
target_plane = plane_state->plane;
break;
}
}
if (target_plane)
weston_view_move_to_plane(ev, &target_plane->base);
else
weston_view_move_to_plane(ev, primary);
if (!target_plane ||
target_plane->type == WDRM_PLANE_TYPE_CURSOR) {
/* cursor plane & renderer involve a copy */
ev->psf_flags = 0;
} else {
/* All other planes are a direct scanout of a
* single client buffer.
*/
ev->psf_flags = WP_PRESENTATION_FEEDBACK_KIND_ZERO_COPY;
}
}
/* We rely on output->cursor_view being both an accurate reflection of
* the cursor plane's state, but also being maintained across repaints
* to avoid unnecessary damage uploads, per the comment in
* drm_output_prepare_cursor_view. In the event that we go from having
* a cursor view to not having a cursor view, we need to clear it. */
if (output->cursor_view) {
plane_state =
drm_output_state_get_existing_plane(state,
output->cursor_plane);
if (!plane_state || !plane_state->fb)
output->cursor_view = NULL;
}
}
/*
* Get the aspect-ratio from drmModeModeInfo mode flags.
*
* @param drm_mode_flags- flags from drmModeModeInfo structure.
* @returns aspect-ratio as encoded in enum 'weston_mode_aspect_ratio'.
*/
static enum weston_mode_aspect_ratio
drm_to_weston_mode_aspect_ratio(uint32_t drm_mode_flags)
{
return (drm_mode_flags & DRM_MODE_FLAG_PIC_AR_MASK) >>
DRM_MODE_FLAG_PIC_AR_BITS_POS;
}
static const char *
aspect_ratio_to_string(enum weston_mode_aspect_ratio ratio)
{
if (ratio < 0 || ratio >= ARRAY_LENGTH(aspect_ratio_as_string) ||
!aspect_ratio_as_string[ratio])
return " (unknown aspect ratio)";
return aspect_ratio_as_string[ratio];
}
/**
* Find the closest-matching mode for a given target
*
* Given a target mode, find the most suitable mode amongst the output's
* current mode list to use, preferring the current mode if possible, to
* avoid an expensive mode switch.
*
* @param output DRM output
* @param target_mode Mode to attempt to match
* @returns Pointer to a mode from the output's mode list
*/
static struct drm_mode *
choose_mode (struct drm_output *output, struct weston_mode *target_mode)
{
struct drm_mode *tmp_mode = NULL, *mode_fall_back = NULL, *mode;
enum weston_mode_aspect_ratio src_aspect = WESTON_MODE_PIC_AR_NONE;
enum weston_mode_aspect_ratio target_aspect = WESTON_MODE_PIC_AR_NONE;
struct drm_backend *b;
b = to_drm_backend(output->base.compositor);
target_aspect = target_mode->aspect_ratio;
src_aspect = output->base.current_mode->aspect_ratio;
if (output->base.current_mode->width == target_mode->width &&
output->base.current_mode->height == target_mode->height &&
(output->base.current_mode->refresh == target_mode->refresh ||
target_mode->refresh == 0)) {
if (!b->aspect_ratio_supported || src_aspect == target_aspect)
return to_drm_mode(output->base.current_mode);
}
wl_list_for_each(mode, &output->base.mode_list, base.link) {
src_aspect = mode->base.aspect_ratio;
if (mode->mode_info.hdisplay == target_mode->width &&
mode->mode_info.vdisplay == target_mode->height) {
if (mode->base.refresh == target_mode->refresh ||
target_mode->refresh == 0) {
if (!b->aspect_ratio_supported ||
src_aspect == target_aspect)
return mode;
else if (!mode_fall_back)
mode_fall_back = mode;
} else if (!tmp_mode) {
tmp_mode = mode;
}
}
}
if (mode_fall_back)
return mode_fall_back;
return tmp_mode;
}
static int
drm_output_init_pixman(struct drm_output *output, struct drm_backend *b);
static void
drm_output_fini_pixman(struct drm_output *output);
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_OPENGL)
static int
drm_output_init_egl(struct drm_output *output, struct drm_backend *b);
static void
drm_output_fini_egl(struct drm_output *output);
#endif
#if defined(ENABLE_IMXG2D)
static int
drm_output_init_g2d(struct drm_output *output, struct drm_backend *b);
static void
drm_output_fini_g2d(struct drm_output *output);
#endif
#endif
static int
drm_output_switch_mode(struct weston_output *output_base, struct weston_mode *mode)
{
struct drm_output *output = to_drm_output(output_base);
struct drm_backend *b = to_drm_backend(output_base->compositor);
struct drm_mode *drm_mode = choose_mode(output, mode);
if (!drm_mode) {
weston_log("%s: invalid resolution %dx%d\n",
output_base->name, mode->width, mode->height);
return -1;
}
if (&drm_mode->base == output->base.current_mode)
return 0;
output->base.current_mode->flags = 0;
output->base.current_mode = &drm_mode->base;
output->base.current_mode->flags =
WL_OUTPUT_MODE_CURRENT | WL_OUTPUT_MODE_PREFERRED;
/* XXX: This drops our current buffer too early, before we've started
* displaying it. Ideally this should be much more atomic and
* integrated with a full repaint cycle, rather than doing a
* sledgehammer modeswitch first, and only later showing new
* content.
*/
b->state_invalid = true;
if (b->use_pixman) {
drm_output_fini_pixman(output);
if (drm_output_init_pixman(output, b) < 0) {
weston_log("failed to init output pixman state with "
"new mode\n");
return -1;
}
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_IMXG2D)
} else if (b->use_g2d) {
drm_output_fini_g2d(output);
if (drm_output_init_g2d(output, b) < 0) {
weston_log("failed to init output g2d state with "
"new mode\n");
return -1;
}
#endif
#if defined(ENABLE_OPENGL)
} else {
drm_output_fini_egl(output);
if (drm_output_init_egl(output, b) < 0) {
weston_log("failed to init output egl state with "
"new mode");
return -1;
}
#endif
#endif
}
return 0;
}
static int
on_drm_input(int fd, uint32_t mask, void *data)
{
#ifdef HAVE_DRM_ATOMIC
struct drm_backend *b = data;
#endif
drmEventContext evctx;
memset(&evctx, 0, sizeof evctx);
#ifndef HAVE_DRM_ATOMIC
evctx.version = 2;
#else
evctx.version = 3;
if (b->atomic_modeset)
evctx.page_flip_handler2 = atomic_flip_handler;
else
#endif
evctx.page_flip_handler = page_flip_handler;
evctx.vblank_handler = vblank_handler;
drmHandleEvent(fd, &evctx);
return 1;
}
static int
init_kms_caps(struct drm_backend *b)
{
uint64_t cap;
int ret;
clockid_t clk_id;
weston_log("using %s\n", b->drm.filename);
ret = drmGetCap(b->drm.fd, DRM_CAP_TIMESTAMP_MONOTONIC, &cap);
if (ret == 0 && cap == 1)
clk_id = CLOCK_MONOTONIC;
else
clk_id = CLOCK_REALTIME;
if (weston_compositor_set_presentation_clock(b->compositor, clk_id) < 0) {
weston_log("Error: failed to set presentation clock %d.\n",
clk_id);
return -1;
}
ret = drmGetCap(b->drm.fd, DRM_CAP_CURSOR_WIDTH, &cap);
if (ret == 0)
b->cursor_width = cap;
else
b->cursor_width = 64;
ret = drmGetCap(b->drm.fd, DRM_CAP_CURSOR_HEIGHT, &cap);
if (ret == 0)
b->cursor_height = cap;
else
b->cursor_height = 64;
if (!getenv("WESTON_DISABLE_UNIVERSAL_PLANES")) {
ret = drmSetClientCap(b->drm.fd, DRM_CLIENT_CAP_UNIVERSAL_PLANES, 1);
b->universal_planes = (ret == 0);
}
weston_log("DRM: %s universal planes\n",
b->universal_planes ? "supports" : "does not support");
#ifdef HAVE_DRM_ATOMIC
if (b->universal_planes && !getenv("WESTON_DISABLE_ATOMIC")) {
ret = drmGetCap(b->drm.fd, DRM_CAP_CRTC_IN_VBLANK_EVENT, &cap);
if (ret != 0)
cap = 0;
ret = drmSetClientCap(b->drm.fd, DRM_CLIENT_CAP_ATOMIC, 1);
b->atomic_modeset = ((ret == 0) && (cap == 1));
}
#endif
weston_log("DRM: %s atomic modesetting\n",
b->atomic_modeset ? "supports" : "does not support");
/*
* KMS support for hardware planes cannot properly synchronize
* without nuclear page flip. Without nuclear/atomic, hw plane
* and cursor plane updates would either tear or cause extra
* waits for vblanks which means dropping the compositor framerate
* to a fraction. For cursors, it's not so bad, so they are
* enabled.
*/
if (!b->atomic_modeset)
b->sprites_are_broken = 1;
ret = drmSetClientCap(b->drm.fd, DRM_CLIENT_CAP_ASPECT_RATIO, 1);
b->aspect_ratio_supported = (ret == 0);
weston_log("DRM: %s picture aspect ratio\n",
b->aspect_ratio_supported ? "supports" : "does not support");
return 0;
}
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
static struct gbm_device *
create_gbm_device(int fd)
{
struct gbm_device *gbm;
gl_renderer = weston_load_module("gl-renderer.so",
"gl_renderer_interface");
if (!gl_renderer)
return NULL;
/* GBM will load a dri driver, but even though they need symbols from
* libglapi, in some version of Mesa they are not linked to it. Since
* only the gl-renderer module links to it, the call above won't make
* these symbols globally available, and loading the DRI driver fails.
* Workaround this by dlopen()'ing libglapi with RTLD_GLOBAL. */
dlopen("libglapi.so.0", RTLD_LAZY | RTLD_GLOBAL);
gbm = gbm_create_device(fd);
return gbm;
}
/* When initializing EGL, if the preferred buffer format isn't available
* we may be able to substitute an ARGB format for an XRGB one.
*
* This returns 0 if substitution isn't possible, but 0 might be a
* legitimate format for other EGL platforms, so the caller is
* responsible for checking for 0 before calling gl_renderer->create().
*
* This works around https://bugs.freedesktop.org/show_bug.cgi?id=89689
* but it's entirely possible we'll see this again on other implementations.
*/
static int
fallback_format_for(uint32_t format)
{
switch (format) {
case GBM_FORMAT_XRGB8888:
return GBM_FORMAT_ARGB8888;
case GBM_FORMAT_XRGB2101010:
return GBM_FORMAT_ARGB2101010;
default:
return 0;
}
}
static int
drm_backend_create_gl_renderer(struct drm_backend *b)
{
EGLint format[3] = {
b->gbm_format,
fallback_format_for(b->gbm_format),
0,
};
int n_formats = 2;
if (format[1])
n_formats = 3;
if (gl_renderer->display_create(b->compositor,
EGL_PLATFORM_GBM_KHR,
(void *)b->gbm,
NULL,
gl_renderer->opaque_attribs,
format,
n_formats) < 0) {
return -1;
}
return 0;
}
static int
init_egl(struct drm_backend *b)
{
b->gbm = create_gbm_device(b->drm.fd);
if (!b->gbm)
return -1;
if (drm_backend_create_gl_renderer(b) < 0) {
gbm_device_destroy(b->gbm);
return -1;
}
return 0;
}
#endif
static int
init_pixman(struct drm_backend *b)
{
return pixman_renderer_init(b->compositor);
}
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
static int
drm_backend_create_g2d_renderer(struct drm_backend *b)
{
if (g2d_renderer->drm_display_create(b->compositor,
(void *)b->gbm) < 0) {
return -1;
}
return 0;
}
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
static int
init_g2d(struct drm_backend *b)
{
g2d_renderer = weston_load_module("g2d-renderer.so",
"g2d_renderer_interface");
if (!g2d_renderer) {
weston_log("Could not load g2d renderer\n");
return -1;
}
b->gbm = gbm_create_device(b->drm.fd);
if (!b->gbm)
return -1;
if (drm_backend_create_g2d_renderer(b) < 0) {
gbm_device_destroy(b->gbm);
return -1;
}
return 0;
}
#endif
#ifdef HAVE_DRM_FORMATS_BLOB
static inline uint32_t *
formats_ptr(struct drm_format_modifier_blob *blob)
{
return (uint32_t *)(((char *)blob) + blob->formats_offset);
}
static inline struct drm_format_modifier *
modifiers_ptr(struct drm_format_modifier_blob *blob)
{
return (struct drm_format_modifier *)
(((char *)blob) + blob->modifiers_offset);
}
#endif
/**
* Populates the plane's formats array, using either the IN_FORMATS blob
* property (if available), or the plane's format list if not.
*/
static int
drm_plane_populate_formats(struct drm_plane *plane, const drmModePlane *kplane,
const drmModeObjectProperties *props)
{
unsigned i;
#ifdef HAVE_DRM_FORMATS_BLOB
drmModePropertyBlobRes *blob;
struct drm_format_modifier_blob *fmt_mod_blob;
struct drm_format_modifier *blob_modifiers;
uint32_t *blob_formats;
uint32_t blob_id;
blob_id = drm_property_get_value(&plane->props[WDRM_PLANE_IN_FORMATS],
props,
0);
if (blob_id == 0)
goto fallback;
blob = drmModeGetPropertyBlob(plane->backend->drm.fd, blob_id);
if (!blob)
goto fallback;
fmt_mod_blob = blob->data;
blob_formats = formats_ptr(fmt_mod_blob);
blob_modifiers = modifiers_ptr(fmt_mod_blob);
if (plane->count_formats != fmt_mod_blob->count_formats) {
weston_log("DRM backend: format count differs between "
"plane (%d) and IN_FORMATS (%d)\n",
plane->count_formats,
fmt_mod_blob->count_formats);
weston_log("This represents a kernel bug; Weston is "
"unable to continue.\n");
abort();
}
for (i = 0; i < fmt_mod_blob->count_formats; i++) {
uint32_t count_modifiers = 0;
uint64_t *modifiers = NULL;
unsigned j;
for (j = 0; j < fmt_mod_blob->count_modifiers; j++) {
struct drm_format_modifier *mod = &blob_modifiers[j];
if ((i < mod->offset) || (i > mod->offset + 63))
continue;
if (!(mod->formats & (1 << (i - mod->offset))))
continue;
modifiers = realloc(modifiers,
(count_modifiers + 1) *
sizeof(modifiers[0]));
assert(modifiers);
modifiers[count_modifiers++] = mod->modifier;
}
plane->formats[i].format = blob_formats[i];
plane->formats[i].modifiers = modifiers;
plane->formats[i].count_modifiers = count_modifiers;
}
drmModeFreePropertyBlob(blob);
return 0;
fallback:
#endif
/* No IN_FORMATS blob available, so just use the old. */
assert(plane->count_formats == kplane->count_formats);
for (i = 0; i < kplane->count_formats; i++)
plane->formats[i].format = kplane->formats[i];
return 0;
}
/**
* Create a drm_plane for a hardware plane
*
* Creates one drm_plane structure for a hardware plane, and initialises its
* properties and formats.
*
* In the absence of universal plane support, where KMS does not explicitly
* expose the primary and cursor planes to userspace, this may also create
* an 'internal' plane for internal management.
*
* This function does not add the plane to the list of usable planes in Weston
* itself; the caller is responsible for this.
*
* Call drm_plane_destroy to clean up the plane.
*
* @sa drm_output_find_special_plane
* @param b DRM compositor backend
* @param kplane DRM plane to create, or NULL if creating internal plane
* @param output Output to create internal plane for, or NULL
* @param type Type to use when creating internal plane, or invalid
* @param format Format to use for internal planes, or 0
*/
static struct drm_plane *
drm_plane_create(struct drm_backend *b, const drmModePlane *kplane,
struct drm_output *output, enum wdrm_plane_type type,
uint32_t format)
{
struct drm_plane *plane;
drmModeObjectProperties *props;
uint32_t num_formats = (kplane) ? kplane->count_formats : 1;
plane = zalloc(sizeof(*plane) +
(sizeof(plane->formats[0]) * num_formats));
if (!plane) {
weston_log("%s: out of memory\n", __func__);
return NULL;
}
plane->backend = b;
plane->count_formats = num_formats;
plane->state_cur = drm_plane_state_alloc(NULL, plane);
plane->state_cur->complete = true;
if (kplane) {
plane->possible_crtcs = kplane->possible_crtcs;
plane->plane_id = kplane->plane_id;
props = drmModeObjectGetProperties(b->drm.fd, kplane->plane_id,
DRM_MODE_OBJECT_PLANE);
if (!props) {
weston_log("couldn't get plane properties\n");
goto err;
}
drm_property_info_populate(b, plane_props, plane->props,
WDRM_PLANE__COUNT, props);
plane->type =
drm_property_get_value(&plane->props[WDRM_PLANE_TYPE],
props,
WDRM_PLANE_TYPE__COUNT);
if (drm_plane_populate_formats(plane, kplane, props) < 0) {
drmModeFreeObjectProperties(props);
goto err;
}
drmModeFreeObjectProperties(props);
}
else {
plane->possible_crtcs = (1 << output->pipe);
plane->plane_id = 0;
plane->count_formats = 1;
plane->formats[0].format = format;
plane->type = type;
}
if (plane->type == WDRM_PLANE_TYPE__COUNT)
goto err_props;
/* With universal planes, everything is a DRM plane; without
* universal planes, the only DRM planes are overlay planes.
* Everything else is a fake plane. */
if (b->universal_planes) {
assert(kplane);
} else {
if (kplane)
assert(plane->type == WDRM_PLANE_TYPE_OVERLAY);
else
assert(plane->type != WDRM_PLANE_TYPE_OVERLAY &&
output);
}
weston_plane_init(&plane->base, b->compositor, 0, 0);
wl_list_insert(&b->plane_list, &plane->link);
return plane;
err_props:
drm_property_info_free(plane->props, WDRM_PLANE__COUNT);
err:
drm_plane_state_free(plane->state_cur, true);
free(plane);
return NULL;
}
/**
* Find, or create, a special-purpose plane
*
* Primary and cursor planes are a special case, in that before universal
* planes, they are driven by non-plane API calls. Without universal plane
* support, the only way to configure a primary plane is via drmModeSetCrtc,
* and the only way to configure a cursor plane is drmModeSetCursor2.
*
* Although they may actually be regular planes in the hardware, without
* universal plane support, these planes are not actually exposed to
* userspace in the regular plane list.
*
* However, for ease of internal tracking, we want to manage all planes
* through the same drm_plane structures. Therefore, when we are running
* without universal plane support, we create fake drm_plane structures
* to track these planes.
*
* @param b DRM backend
* @param output Output to use for plane
* @param type Type of plane
*/
static struct drm_plane *
drm_output_find_special_plane(struct drm_backend *b, struct drm_output *output,
enum wdrm_plane_type type)
{
struct drm_plane *plane;
if (!b->universal_planes) {
uint32_t format;
switch (type) {
case WDRM_PLANE_TYPE_CURSOR:
format = GBM_FORMAT_ARGB8888;
break;
case WDRM_PLANE_TYPE_PRIMARY:
/* We don't know what formats the primary plane supports
* before universal planes, so we just assume that the
* GBM format works; however, this isn't set until after
* the output is created. */
format = 0;
break;
default:
assert(!"invalid type in drm_output_find_special_plane");
break;
}
return drm_plane_create(b, NULL, output, type, format);
}
wl_list_for_each(plane, &b->plane_list, link) {
struct drm_output *tmp;
bool found_elsewhere = false;
if (plane->type != type)
continue;
if (!drm_plane_is_available(plane, output))
continue;
/* On some platforms, primary/cursor planes can roam
* between different CRTCs, so make sure we don't claim the
* same plane for two outputs. */
wl_list_for_each(tmp, &b->compositor->output_list,
base.link) {
if (tmp->cursor_plane == plane ||
tmp->scanout_plane == plane) {
found_elsewhere = true;
break;
}
}
if (found_elsewhere)
continue;
plane->possible_crtcs = (1 << output->pipe);
return plane;
}
return NULL;
}
/**
* Destroy one DRM plane
*
* Destroy a DRM plane, removing it from screen and releasing its retained
* buffers in the process. The counterpart to drm_plane_create.
*
* @param plane Plane to deallocate (will be freed)
*/
static void
drm_plane_destroy(struct drm_plane *plane)
{
if (plane->type == WDRM_PLANE_TYPE_OVERLAY)
drmModeSetPlane(plane->backend->drm.fd, plane->plane_id,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0);
drm_plane_state_free(plane->state_cur, true);
drm_property_info_free(plane->props, WDRM_PLANE__COUNT);
weston_plane_release(&plane->base);
wl_list_remove(&plane->link);
free(plane);
}
/**
* Initialise sprites (overlay planes)
*
* Walk the list of provided DRM planes, and add overlay planes.
*
* Call destroy_sprites to free these planes.
*
* @param b DRM compositor backend
*/
static void
create_sprites(struct drm_backend *b)
{
drmModePlaneRes *kplane_res;
drmModePlane *kplane;
struct drm_plane *drm_plane;
uint32_t i;
kplane_res = drmModeGetPlaneResources(b->drm.fd);
if (!kplane_res) {
weston_log("failed to get plane resources: %s\n",
strerror(errno));
return;
}
for (i = 0; i < kplane_res->count_planes; i++) {
kplane = drmModeGetPlane(b->drm.fd, kplane_res->planes[i]);
if (!kplane)
continue;
drm_plane = drm_plane_create(b, kplane, NULL,
WDRM_PLANE_TYPE__COUNT, 0);
drmModeFreePlane(kplane);
if (!drm_plane)
continue;
if (drm_plane->type == WDRM_PLANE_TYPE_OVERLAY)
weston_compositor_stack_plane(b->compositor,
&drm_plane->base,
&b->compositor->primary_plane);
}
drmModeFreePlaneResources(kplane_res);
}
/**
* Clean up sprites (overlay planes)
*
* The counterpart to create_sprites.
*
* @param b DRM compositor backend
*/
static void
destroy_sprites(struct drm_backend *b)
{
struct drm_plane *plane, *next;
wl_list_for_each_safe(plane, next, &b->plane_list, link)
drm_plane_destroy(plane);
}
static uint32_t
drm_refresh_rate_mHz(const drmModeModeInfo *info)
{
uint64_t refresh;
/* Calculate higher precision (mHz) refresh rate */
refresh = (info->clock * 1000000LL / info->htotal +
info->vtotal / 2) / info->vtotal;
if (info->flags & DRM_MODE_FLAG_INTERLACE)
refresh *= 2;
if (info->flags & DRM_MODE_FLAG_DBLSCAN)
refresh /= 2;
if (info->vscan > 1)
refresh /= info->vscan;
return refresh;
}
/**
* Add a mode to output's mode list
*
* Copy the supplied DRM mode into a Weston mode structure, and add it to the
* output's mode list.
*
* @param output DRM output to add mode to
* @param info DRM mode structure to add
* @returns Newly-allocated Weston/DRM mode structure
*/
static struct drm_mode *
drm_output_add_mode(struct drm_output *output, const drmModeModeInfo *info)
{
struct drm_mode *mode;
mode = malloc(sizeof *mode);
if (mode == NULL)
return NULL;
mode->base.flags = 0;
mode->base.width = info->hdisplay;
mode->base.height = info->vdisplay;
mode->base.refresh = drm_refresh_rate_mHz(info);
mode->mode_info = *info;
mode->blob_id = 0;
if (info->type & DRM_MODE_TYPE_PREFERRED)
mode->base.flags |= WL_OUTPUT_MODE_PREFERRED;
mode->base.aspect_ratio = drm_to_weston_mode_aspect_ratio(info->flags);
wl_list_insert(output->base.mode_list.prev, &mode->base.link);
return mode;
}
/**
* Destroys a mode, and removes it from the list.
*/
static void
drm_output_destroy_mode(struct drm_backend *backend, struct drm_mode *mode)
{
if (mode->blob_id)
drmModeDestroyPropertyBlob(backend->drm.fd, mode->blob_id);
wl_list_remove(&mode->base.link);
free(mode);
}
/** Destroy a list of drm_modes
*
* @param backend The backend for releasing mode property blobs.
* @param mode_list The list linked by drm_mode::base.link.
*/
static void
drm_mode_list_destroy(struct drm_backend *backend, struct wl_list *mode_list)
{
struct drm_mode *mode, *next;
wl_list_for_each_safe(mode, next, mode_list, base.link)
drm_output_destroy_mode(backend, mode);
}
static int
drm_subpixel_to_wayland(int drm_value)
{
switch (drm_value) {
default:
case DRM_MODE_SUBPIXEL_UNKNOWN:
return WL_OUTPUT_SUBPIXEL_UNKNOWN;
case DRM_MODE_SUBPIXEL_NONE:
return WL_OUTPUT_SUBPIXEL_NONE;
case DRM_MODE_SUBPIXEL_HORIZONTAL_RGB:
return WL_OUTPUT_SUBPIXEL_HORIZONTAL_RGB;
case DRM_MODE_SUBPIXEL_HORIZONTAL_BGR:
return WL_OUTPUT_SUBPIXEL_HORIZONTAL_BGR;
case DRM_MODE_SUBPIXEL_VERTICAL_RGB:
return WL_OUTPUT_SUBPIXEL_VERTICAL_RGB;
case DRM_MODE_SUBPIXEL_VERTICAL_BGR:
return WL_OUTPUT_SUBPIXEL_VERTICAL_BGR;
}
}
/* returns a value between 0-255 range, where higher is brighter */
static uint32_t
drm_get_backlight(struct drm_head *head)
{
long brightness, max_brightness, norm;
brightness = backlight_get_brightness(head->backlight);
max_brightness = backlight_get_max_brightness(head->backlight);
/* convert it on a scale of 0 to 255 */
norm = (brightness * 255)/(max_brightness);
return (uint32_t) norm;
}
/* values accepted are between 0-255 range */
static void
drm_set_backlight(struct weston_output *output_base, uint32_t value)
{
struct drm_output *output = to_drm_output(output_base);
struct drm_head *head;
long max_brightness, new_brightness;
if (value > 255)
return;
wl_list_for_each(head, &output->base.head_list, base.output_link) {
if (!head->backlight)
return;
max_brightness = backlight_get_max_brightness(head->backlight);
/* get denormalized value */
new_brightness = (value * max_brightness) / 255;
backlight_set_brightness(head->backlight, new_brightness);
}
}
static void
drm_output_init_backlight(struct drm_output *output)
{
struct weston_head *base;
struct drm_head *head;
output->base.set_backlight = NULL;
wl_list_for_each(base, &output->base.head_list, output_link) {
head = to_drm_head(base);
if (head->backlight) {
weston_log("Initialized backlight for head '%s', device %s\n",
head->base.name, head->backlight->path);
if (!output->base.set_backlight) {
output->base.set_backlight = drm_set_backlight;
output->base.backlight_current =
drm_get_backlight(head);
}
}
}
if (!output->base.set_backlight) {
weston_log("No backlight control for output '%s'\n",
output->base.name);
}
}
/**
* Power output on or off
*
* The DPMS/power level of an output is used to switch it on or off. This
* is DRM's hook for doing so, which can called either as part of repaint,
* or independently of the repaint loop.
*
* If we are called as part of repaint, we simply set the relevant bit in
* state and return.
*/
static void
drm_set_dpms(struct weston_output *output_base, enum dpms_enum level)
{
struct drm_output *output = to_drm_output(output_base);
struct drm_backend *b = to_drm_backend(output_base->compositor);
struct drm_pending_state *pending_state = b->repaint_data;
struct drm_output_state *state;
int ret;
if (output->state_cur->dpms == level)
return;
/* If we're being called during the repaint loop, then this is
* simple: discard any previously-generated state, and create a new
* state where we disable everything. When we come to flush, this
* will be applied.
*
* However, we need to be careful: we can be called whilst another
* output is in its repaint cycle (pending_state exists), but our
* output still has an incomplete state application outstanding.
* In that case, we need to wait until that completes. */
if (pending_state && !output->state_last) {
/* The repaint loop already sets DPMS on; we don't need to
* explicitly set it on here, as it will already happen
* whilst applying the repaint state. */
if (level == WESTON_DPMS_ON)
return;
state = drm_pending_state_get_output(pending_state, output);
if (state)
drm_output_state_free(state);
state = drm_output_get_disable_state(pending_state, output);
return;
}
/* As we throw everything away when disabling, just send us back through
* a repaint cycle. */
if (level == WESTON_DPMS_ON) {
if (output->dpms_off_pending)
output->dpms_off_pending = 0;
weston_output_schedule_repaint(output_base);
return;
}
/* If we've already got a request in the pipeline, then we need to
* park our DPMS request until that request has quiesced. */
if (output->state_last) {
output->dpms_off_pending = 1;
return;
}
pending_state = drm_pending_state_alloc(b);
drm_output_get_disable_state(pending_state, output);
ret = drm_pending_state_apply_sync(pending_state);
if (ret != 0)
weston_log("drm_set_dpms: couldn't disable output?\n");
}
static const char * const connector_type_names[] = {
[DRM_MODE_CONNECTOR_Unknown] = "Unknown",
[DRM_MODE_CONNECTOR_VGA] = "VGA",
[DRM_MODE_CONNECTOR_DVII] = "DVI-I",
[DRM_MODE_CONNECTOR_DVID] = "DVI-D",
[DRM_MODE_CONNECTOR_DVIA] = "DVI-A",
[DRM_MODE_CONNECTOR_Composite] = "Composite",
[DRM_MODE_CONNECTOR_SVIDEO] = "SVIDEO",
[DRM_MODE_CONNECTOR_LVDS] = "LVDS",
[DRM_MODE_CONNECTOR_Component] = "Component",
[DRM_MODE_CONNECTOR_9PinDIN] = "DIN",
[DRM_MODE_CONNECTOR_DisplayPort] = "DP",
[DRM_MODE_CONNECTOR_HDMIA] = "HDMI-A",
[DRM_MODE_CONNECTOR_HDMIB] = "HDMI-B",
[DRM_MODE_CONNECTOR_TV] = "TV",
[DRM_MODE_CONNECTOR_eDP] = "eDP",
#ifdef DRM_MODE_CONNECTOR_DSI
[DRM_MODE_CONNECTOR_VIRTUAL] = "Virtual",
[DRM_MODE_CONNECTOR_DSI] = "DSI",
#endif
#ifdef DRM_MODE_CONNECTOR_DPI
[DRM_MODE_CONNECTOR_DPI] = "DPI",
#endif
};
/** Create a name given a DRM connector
*
* \param con The DRM connector whose type and id form the name.
* \return A newly allocate string, or NULL on error. Must be free()'d
* after use.
*
* The name does not identify the DRM display device.
*/
static char *
make_connector_name(const drmModeConnector *con)
{
char *name;
const char *type_name = NULL;
int ret;
if (con->connector_type < ARRAY_LENGTH(connector_type_names))
type_name = connector_type_names[con->connector_type];
if (!type_name)
type_name = "UNNAMED";
ret = asprintf(&name, "%s-%d", type_name, con->connector_type_id);
if (ret < 0)
return NULL;
return name;
}
static void drm_output_fini_cursor_egl(struct drm_output *output)
{
unsigned int i;
for (i = 0; i < ARRAY_LENGTH(output->gbm_cursor_fb); i++) {
drm_fb_unref(output->gbm_cursor_fb[i]);
output->gbm_cursor_fb[i] = NULL;
}
}
static int
drm_output_init_cursor_egl(struct drm_output *output, struct drm_backend *b)
{
unsigned int i;
/* No point creating cursors if we don't have a plane for them. */
if (!output->cursor_plane)
return 0;
for (i = 0; i < ARRAY_LENGTH(output->gbm_cursor_fb); i++) {
struct gbm_bo *bo;
bo = gbm_bo_create(b->gbm, b->cursor_width, b->cursor_height,
GBM_FORMAT_ARGB8888,
GBM_BO_USE_CURSOR | GBM_BO_USE_WRITE);
if (!bo)
goto err;
output->gbm_cursor_fb[i] =
drm_fb_get_from_bo(bo, b, false, BUFFER_CURSOR);
if (!output->gbm_cursor_fb[i]) {
gbm_bo_destroy(bo);
goto err;
}
}
return 0;
err:
weston_log("cursor buffers unavailable, using gl cursors\n");
b->cursors_are_broken = 1;
drm_output_fini_cursor_egl(output);
return -1;
}
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
/* Init output state that depends on gl or gbm */
static int
drm_output_init_egl(struct drm_output *output, struct drm_backend *b)
{
EGLint format[2] = {
output->gbm_format,
fallback_format_for(output->gbm_format),
};
int n_formats = 1;
struct weston_mode *mode = output->base.current_mode;
struct drm_plane *plane = output->scanout_plane;
unsigned int i;
for (i = 0; i < plane->count_formats; i++) {
if (plane->formats[i].format == output->gbm_format)
break;
}
if (i == plane->count_formats) {
weston_log("format 0x%x not supported by output %s\n",
output->gbm_format, output->base.name);
return -1;
}
#ifdef HAVE_GBM_MODIFIERS
if (plane->formats[i].count_modifiers > 0) {
output->gbm_surface =
gbm_surface_create_with_modifiers(b->gbm,
mode->width,
mode->height,
output->gbm_format,
plane->formats[i].modifiers,
plane->formats[i].count_modifiers);
} else
#endif
{
output->gbm_surface =
gbm_surface_create(b->gbm, mode->width, mode->height,
output->gbm_format,
GBM_BO_USE_RENDERING | GBM_BO_USE_SCANOUT);
}
if (!output->gbm_surface) {
weston_log("failed to create gbm surface\n");
return -1;
}
if (format[1])
n_formats = 2;
if (gl_renderer->output_window_create(&output->base,
(EGLNativeWindowType)output->gbm_surface,
output->gbm_surface,
gl_renderer->opaque_attribs,
format,
n_formats) < 0) {
weston_log("failed to create gl renderer output state\n");
gbm_surface_destroy(output->gbm_surface);
return -1;
}
drm_output_init_cursor_egl(output, b);
return 0;
}
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
static void
drm_output_fini_egl(struct drm_output *output)
{
struct drm_backend *b = to_drm_backend(output->base.compositor);
/* Destroying the GBM surface will destroy all our GBM buffers,
* regardless of refcount. Ensure we destroy them here. */
if (!b->shutting_down &&
output->scanout_plane->state_cur->fb &&
output->scanout_plane->state_cur->fb->type == BUFFER_GBM_SURFACE) {
drm_plane_state_free(output->scanout_plane->state_cur, true);
output->scanout_plane->state_cur =
drm_plane_state_alloc(NULL, output->scanout_plane);
output->scanout_plane->state_cur->complete = true;
}
gl_renderer->output_destroy(&output->base);
gbm_surface_destroy(output->gbm_surface);
drm_output_fini_cursor_egl(output);
}
#endif
static int
drm_output_init_pixman(struct drm_output *output, struct drm_backend *b)
{
int w = output->base.current_mode->width;
int h = output->base.current_mode->height;
uint32_t format = output->gbm_format;
uint32_t pixman_format;
unsigned int i;
uint32_t flags = 0;
switch (format) {
case GBM_FORMAT_XRGB8888:
pixman_format = PIXMAN_x8r8g8b8;
break;
case GBM_FORMAT_RGB565:
pixman_format = PIXMAN_r5g6b5;
break;
default:
weston_log("Unsupported pixman format 0x%x\n", format);
return -1;
}
/* FIXME error checking */
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
output->dumb[i] = drm_fb_create_dumb(b, w, h, format);
if (!output->dumb[i])
goto err;
output->image[i] =
pixman_image_create_bits(pixman_format, w, h,
output->dumb[i]->map,
output->dumb[i]->strides[0]);
if (!output->image[i])
goto err;
}
if (b->use_pixman_shadow)
flags |= PIXMAN_RENDERER_OUTPUT_USE_SHADOW;
if (pixman_renderer_output_create(&output->base, flags) < 0)
goto err;
weston_log("DRM: output %s %s shadow framebuffer.\n", output->base.name,
b->use_pixman_shadow ? "uses" : "does not use");
pixman_region32_init_rect(&output->previous_damage,
output->base.x, output->base.y, output->base.width, output->base.height);
return 0;
err:
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
if (output->dumb[i])
drm_fb_unref(output->dumb[i]);
if (output->image[i])
pixman_image_unref(output->image[i]);
output->dumb[i] = NULL;
output->image[i] = NULL;
}
return -1;
}
static void
drm_output_fini_pixman(struct drm_output *output)
{
struct drm_backend *b = to_drm_backend(output->base.compositor);
unsigned int i;
/* Destroying the Pixman surface will destroy all our buffers,
* regardless of refcount. Ensure we destroy them here. */
if (!b->shutting_down &&
output->scanout_plane->state_cur->fb &&
output->scanout_plane->state_cur->fb->type == BUFFER_PIXMAN_DUMB) {
drm_plane_state_free(output->scanout_plane->state_cur, true);
output->scanout_plane->state_cur =
drm_plane_state_alloc(NULL, output->scanout_plane);
output->scanout_plane->state_cur->complete = true;
}
pixman_renderer_output_destroy(&output->base);
pixman_region32_fini(&output->previous_damage);
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
pixman_image_unref(output->image[i]);
drm_fb_unref(output->dumb[i]);
output->dumb[i] = NULL;
output->image[i] = NULL;
}
}
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
static int
drm_output_init_g2d(struct drm_output *output, struct drm_backend *b)
{
int w = output->base.current_mode->width;
int h = output->base.current_mode->height;
uint32_t format = output->gbm_format;
enum g2d_format g2dFormat;
uint32_t i = 0;
switch (format) {
case GBM_FORMAT_XRGB8888:
g2dFormat = G2D_BGRX8888;
break;
case GBM_FORMAT_RGB565:
g2dFormat = G2D_RGB565;
break;
default:
weston_log("Unsupported pixman format 0x%x\n", format);
return -1;
}
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
struct g2d_surfaceEx* g2dSurface = &(output->g2d_image[i]);
int ret, dmafd = 0;
output->dumb[i] = drm_fb_create_dumb(b, w, h, format);
if (!output->dumb[i])
goto err;
ret = drmPrimeHandleToFD(b->drm.fd, output->dumb[i]->handles[0], DRM_CLOEXEC,
&dmafd);
if(ret < 0)
goto err;
ret = g2d_renderer->create_g2d_image(g2dSurface, g2dFormat,
output->dumb[i]->map,
w, h,
output->dumb[i]->strides[0],
output->dumb[i]->size,
dmafd);
if (ret < 0)
goto err;
}
if (g2d_renderer->drm_output_create(&output->base) < 0)
goto err;
drm_output_init_cursor_egl(output, b);
return 0;
err:
weston_log("drm_output_init_g2d failed.\n");
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
if (output->dumb[i])
drm_fb_unref(output->dumb[i]);
output->dumb[i] = NULL;
}
return -1;
}
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
static void
drm_output_fini_g2d(struct drm_output *output)
{
unsigned int i;
pixman_region32_fini(&output->previous_damage);
for (i = 0; i < ARRAY_LENGTH(output->dumb); i++) {
drm_fb_unref(output->dumb[i]);
output->dumb[i] = NULL;
}
g2d_renderer->output_destroy(&output->base);
}
#endif
static void
edid_parse_string(const uint8_t *data, char text[])
{
int i;
int replaced = 0;
/* this is always 12 bytes, but we can't guarantee it's null
* terminated or not junk. */
strncpy(text, (const char *) data, 12);
/* guarantee our new string is null-terminated */
text[12] = '\0';
/* remove insane chars */
for (i = 0; text[i] != '\0'; i++) {
if (text[i] == '\n' ||
text[i] == '\r') {
text[i] = '\0';
break;
}
}
/* ensure string is printable */
for (i = 0; text[i] != '\0'; i++) {
if (!isprint(text[i])) {
text[i] = '-';
replaced++;
}
}
/* if the string is random junk, ignore the string */
if (replaced > 4)
text[0] = '\0';
}
#define EDID_DESCRIPTOR_ALPHANUMERIC_DATA_STRING 0xfe
#define EDID_DESCRIPTOR_DISPLAY_PRODUCT_NAME 0xfc
#define EDID_DESCRIPTOR_DISPLAY_PRODUCT_SERIAL_NUMBER 0xff
#define EDID_OFFSET_DATA_BLOCKS 0x36
#define EDID_OFFSET_LAST_BLOCK 0x6c
#define EDID_OFFSET_PNPID 0x08
#define EDID_OFFSET_SERIAL 0x0c
static int
edid_parse(struct drm_edid *edid, const uint8_t *data, size_t length)
{
int i;
uint32_t serial_number;
/* check header */
if (length < 128)
return -1;
if (data[0] != 0x00 || data[1] != 0xff)
return -1;
/* decode the PNP ID from three 5 bit words packed into 2 bytes
* /--08--\/--09--\
* 7654321076543210
* |\---/\---/\---/
* R C1 C2 C3 */
edid->pnp_id[0] = 'A' + ((data[EDID_OFFSET_PNPID + 0] & 0x7c) / 4) - 1;
edid->pnp_id[1] = 'A' + ((data[EDID_OFFSET_PNPID + 0] & 0x3) * 8) + ((data[EDID_OFFSET_PNPID + 1] & 0xe0) / 32) - 1;
edid->pnp_id[2] = 'A' + (data[EDID_OFFSET_PNPID + 1] & 0x1f) - 1;
edid->pnp_id[3] = '\0';
/* maybe there isn't a ASCII serial number descriptor, so use this instead */
serial_number = (uint32_t) data[EDID_OFFSET_SERIAL + 0];
serial_number += (uint32_t) data[EDID_OFFSET_SERIAL + 1] * 0x100;
serial_number += (uint32_t) data[EDID_OFFSET_SERIAL + 2] * 0x10000;
serial_number += (uint32_t) data[EDID_OFFSET_SERIAL + 3] * 0x1000000;
if (serial_number > 0)
sprintf(edid->serial_number, "%lu", (unsigned long) serial_number);
/* parse EDID data */
for (i = EDID_OFFSET_DATA_BLOCKS;
i <= EDID_OFFSET_LAST_BLOCK;
i += 18) {
/* ignore pixel clock data */
if (data[i] != 0)
continue;
if (data[i+2] != 0)
continue;
/* any useful blocks? */
if (data[i+3] == EDID_DESCRIPTOR_DISPLAY_PRODUCT_NAME) {
edid_parse_string(&data[i+5],
edid->monitor_name);
} else if (data[i+3] == EDID_DESCRIPTOR_DISPLAY_PRODUCT_SERIAL_NUMBER) {
edid_parse_string(&data[i+5],
edid->serial_number);
} else if (data[i+3] == EDID_DESCRIPTOR_ALPHANUMERIC_DATA_STRING) {
edid_parse_string(&data[i+5],
edid->eisa_id);
}
}
return 0;
}
/** Parse monitor make, model and serial from EDID
*
* \param head The head whose \c drm_edid to fill in.
* \param props The DRM connector properties to get the EDID from.
* \param make[out] The monitor make (PNP ID).
* \param model[out] The monitor model (name).
* \param serial_number[out] The monitor serial number.
*
* Each of \c *make, \c *model and \c *serial_number are set only if the
* information is found in the EDID. The pointers they are set to must not
* be free()'d explicitly, instead they get implicitly freed when the
* \c drm_head is destroyed.
*/
static void
find_and_parse_output_edid(struct drm_head *head,
drmModeObjectPropertiesPtr props,
const char **make,
const char **model,
const char **serial_number)
{
drmModePropertyBlobPtr edid_blob = NULL;
uint32_t blob_id;
int rc;
blob_id =
drm_property_get_value(&head->props_conn[WDRM_CONNECTOR_EDID],
props, 0);
if (!blob_id)
return;
edid_blob = drmModeGetPropertyBlob(head->backend->drm.fd, blob_id);
if (!edid_blob)
return;
rc = edid_parse(&head->edid,
edid_blob->data,
edid_blob->length);
if (!rc) {
if (head->edid.pnp_id[0] != '\0')
*make = head->edid.pnp_id;
if (head->edid.monitor_name[0] != '\0')
*model = head->edid.monitor_name;
if (head->edid.serial_number[0] != '\0')
*serial_number = head->edid.serial_number;
}
drmModeFreePropertyBlob(edid_blob);
}
static int
parse_modeline(const char *s, drmModeModeInfo *mode)
{
char hsync[16];
char vsync[16];
float fclock;
memset(mode, 0, sizeof *mode);
mode->type = DRM_MODE_TYPE_USERDEF;
mode->hskew = 0;
mode->vscan = 0;
mode->vrefresh = 0;
mode->flags = 0;
if (sscanf(s, "%f %hd %hd %hd %hd %hd %hd %hd %hd %15s %15s",
&fclock,
&mode->hdisplay,
&mode->hsync_start,
&mode->hsync_end,
&mode->htotal,
&mode->vdisplay,
&mode->vsync_start,
&mode->vsync_end,
&mode->vtotal, hsync, vsync) != 11)
return -1;
mode->clock = fclock * 1000;
if (strcasecmp(hsync, "+hsync") == 0)
mode->flags |= DRM_MODE_FLAG_PHSYNC;
else if (strcasecmp(hsync, "-hsync") == 0)
mode->flags |= DRM_MODE_FLAG_NHSYNC;
else
return -1;
if (strcasecmp(vsync, "+vsync") == 0)
mode->flags |= DRM_MODE_FLAG_PVSYNC;
else if (strcasecmp(vsync, "-vsync") == 0)
mode->flags |= DRM_MODE_FLAG_NVSYNC;
else
return -1;
snprintf(mode->name, sizeof mode->name, "%dx%d@%.3f",
mode->hdisplay, mode->vdisplay, fclock);
return 0;
}
static void
setup_output_seat_constraint(struct drm_backend *b,
struct weston_output *output,
const char *s)
{
if (strcmp(s, "") != 0) {
struct weston_pointer *pointer;
struct udev_seat *seat;
seat = udev_seat_get_named(&b->input, s);
if (!seat)
return;
seat->base.output = output;
pointer = weston_seat_get_pointer(&seat->base);
if (pointer)
weston_pointer_clamp(pointer,
&pointer->x,
&pointer->y);
}
}
static int
drm_output_attach_head(struct weston_output *output_base,
struct weston_head *head_base)
{
struct drm_backend *b = to_drm_backend(output_base->compositor);
if (wl_list_length(&output_base->head_list) >= MAX_CLONED_CONNECTORS)
return -1;
if (!output_base->enabled)
return 0;
/* XXX: ensure the configuration will work.
* This is actually impossible without major infrastructure
* work. */
/* Need to go through modeset to add connectors. */
/* XXX: Ideally we'd do this per-output, not globally. */
/* XXX: Doing it globally, what guarantees another output's update
* will not clear the flag before this output is updated?
*/
b->state_invalid = true;
weston_output_schedule_repaint(output_base);
return 0;
}
static void
drm_output_detach_head(struct weston_output *output_base,
struct weston_head *head_base)
{
struct drm_backend *b = to_drm_backend(output_base->compositor);
if (!output_base->enabled)
return;
/* Need to go through modeset to drop connectors that should no longer
* be driven. */
/* XXX: Ideally we'd do this per-output, not globally. */
b->state_invalid = true;
weston_output_schedule_repaint(output_base);
}
static int
parse_gbm_format(const char *s, uint32_t default_value, uint32_t *gbm_format)
{
int ret = 0;
if (s == NULL)
*gbm_format = default_value;
else if (strcmp(s, "xrgb8888") == 0)
*gbm_format = GBM_FORMAT_XRGB8888;
else if (strcmp(s, "argb8888") == 0)
*gbm_format = GBM_FORMAT_ARGB8888;
else if (strcmp(s, "rgb565") == 0)
*gbm_format = GBM_FORMAT_RGB565;
else if (strcmp(s, "xrgb2101010") == 0)
*gbm_format = GBM_FORMAT_XRGB2101010;
else {
weston_log("fatal: unrecognized pixel format: %s\n", s);
ret = -1;
}
return ret;
}
static uint32_t
u32distance(uint32_t a, uint32_t b)
{
if (a < b)
return b - a;
else
return a - b;
}
/** Choose equivalent mode
*
* If the two modes are not equivalent, return NULL.
* Otherwise return the mode that is more likely to work in place of both.
*
* None of the fuzzy matching criteria in this function have any justification.
*
* typedef struct _drmModeModeInfo {
* uint32_t clock;
* uint16_t hdisplay, hsync_start, hsync_end, htotal, hskew;
* uint16_t vdisplay, vsync_start, vsync_end, vtotal, vscan;
*
* uint32_t vrefresh;
*
* uint32_t flags;
* uint32_t type;
* char name[DRM_DISPLAY_MODE_LEN];
* } drmModeModeInfo, *drmModeModeInfoPtr;
*/
static const drmModeModeInfo *
drm_mode_pick_equivalent(const drmModeModeInfo *a, const drmModeModeInfo *b)
{
uint32_t refresh_a, refresh_b;
if (a->hdisplay != b->hdisplay || a->vdisplay != b->vdisplay)
return NULL;
if (a->flags != b->flags)
return NULL;
/* kHz */
if (u32distance(a->clock, b->clock) > 500)
return NULL;
refresh_a = drm_refresh_rate_mHz(a);
refresh_b = drm_refresh_rate_mHz(b);
if (u32distance(refresh_a, refresh_b) > 50)
return NULL;
if ((a->type ^ b->type) & DRM_MODE_TYPE_PREFERRED) {
if (a->type & DRM_MODE_TYPE_PREFERRED)
return a;
else
return b;
}
return a;
}
/* If the given mode info is not already in the list, add it.
* If it is in the list, either keep the existing or replace it,
* depending on which one is "better".
*/
static int
drm_output_try_add_mode(struct drm_output *output, const drmModeModeInfo *info)
{
struct weston_mode *base;
struct drm_mode *mode;
struct drm_backend *backend;
const drmModeModeInfo *chosen = NULL;
assert(info);
wl_list_for_each(base, &output->base.mode_list, link) {
mode = to_drm_mode(base);
chosen = drm_mode_pick_equivalent(&mode->mode_info, info);
if (chosen)
break;
}
if (chosen == info) {
backend = to_drm_backend(output->base.compositor);
drm_output_destroy_mode(backend, mode);
chosen = NULL;
}
if (!chosen) {
mode = drm_output_add_mode(output, info);
if (!mode)
return -1;
}
/* else { the equivalent mode is already in the list } */
return 0;
}
/** Rewrite the output's mode list
*
* @param output The output.
* @return 0 on success, -1 on failure.
*
* Destroy all existing modes in the list, and reconstruct a new list from
* scratch, based on the currently attached heads.
*
* On failure the output's mode list may contain some modes.
*/
static int
drm_output_update_modelist_from_heads(struct drm_output *output)
{
struct drm_backend *backend = to_drm_backend(output->base.compositor);
struct weston_head *head_base;
struct drm_head *head;
int i;
int ret;
assert(!output->base.enabled);
drm_mode_list_destroy(backend, &output->base.mode_list);
wl_list_for_each(head_base, &output->base.head_list, output_link) {
head = to_drm_head(head_base);
for (i = 0; i < head->connector->count_modes; i++) {
ret = drm_output_try_add_mode(output,
&head->connector->modes[i]);
if (ret < 0)
return -1;
}
}
return 0;
}
/**
* Choose suitable mode for an output
*
* Find the most suitable mode to use for initial setup (or reconfiguration on
* hotplug etc) for a DRM output.
*
* @param output DRM output to choose mode for
* @param kind Strategy and preference to use when choosing mode
* @param width Desired width for this output
* @param height Desired height for this output
* @param current_mode Mode currently being displayed on this output
* @param modeline Manually-entered mode (may be NULL)
* @returns A mode from the output's mode list, or NULL if none available
*/
static struct drm_mode *
drm_output_choose_initial_mode(struct drm_backend *backend,
struct drm_output *output,
enum weston_drm_backend_output_mode mode,
const char *modeline,
const drmModeModeInfo *current_mode)
{
struct drm_mode *preferred = NULL;
struct drm_mode *current = NULL;
struct drm_mode *configured = NULL;
struct drm_mode *config_fall_back = NULL;
struct drm_mode *best = NULL;
struct drm_mode *drm_mode;
drmModeModeInfo drm_modeline;
int32_t width = 0;
int32_t height = 0;
uint32_t refresh = 0;
uint32_t aspect_width = 0;
uint32_t aspect_height = 0;
enum weston_mode_aspect_ratio aspect_ratio = WESTON_MODE_PIC_AR_NONE;
int n;
if (mode == WESTON_DRM_BACKEND_OUTPUT_PREFERRED && modeline) {
n = sscanf(modeline, "%dx%d@%d %u:%u", &width, &height,
&refresh, &aspect_width, &aspect_height);
if (backend->aspect_ratio_supported && n == 5) {
if (aspect_width == 4 && aspect_height == 3)
aspect_ratio = WESTON_MODE_PIC_AR_4_3;
else if (aspect_width == 16 && aspect_height == 9)
aspect_ratio = WESTON_MODE_PIC_AR_16_9;
else if (aspect_width == 64 && aspect_height == 27)
aspect_ratio = WESTON_MODE_PIC_AR_64_27;
else if (aspect_width == 256 && aspect_height == 135)
aspect_ratio = WESTON_MODE_PIC_AR_256_135;
else
weston_log("Invalid modeline \"%s\" for output %s\n",
modeline, output->base.name);
}
if (n != 2 && n != 3 && n != 5) {
width = -1;
if (parse_modeline(modeline, &drm_modeline) == 0) {
configured = drm_output_add_mode(output, &drm_modeline);
if (!configured)
return NULL;
} else {
weston_log("Invalid modeline \"%s\" for output %s\n",
modeline, output->base.name);
}
}
}
wl_list_for_each_reverse(drm_mode, &output->base.mode_list, base.link) {
if (width == drm_mode->base.width &&
height == drm_mode->base.height &&
(refresh == 0 || refresh == drm_mode->mode_info.vrefresh)) {
if (!backend->aspect_ratio_supported ||
aspect_ratio == drm_mode->base.aspect_ratio)
configured = drm_mode;
else
config_fall_back = drm_mode;
}
if (memcmp(current_mode, &drm_mode->mode_info,
sizeof *current_mode) == 0)
current = drm_mode;
if (drm_mode->base.flags & WL_OUTPUT_MODE_PREFERRED)
preferred = drm_mode;
best = drm_mode;
}
if (current == NULL && current_mode->clock != 0) {
current = drm_output_add_mode(output, current_mode);
if (!current)
return NULL;
}
if (mode == WESTON_DRM_BACKEND_OUTPUT_CURRENT)
configured = current;
if (configured)
return configured;
if (config_fall_back)
return config_fall_back;
if (preferred)
return preferred;
if (current)
return current;
if (best)
return best;
weston_log("no available modes for %s\n", output->base.name);
return NULL;
}
static int
drm_head_read_current_setup(struct drm_head *head, struct drm_backend *backend)
{
int drm_fd = backend->drm.fd;
drmModeEncoder *encoder;
drmModeCrtc *crtc;
/* Get the current mode on the crtc that's currently driving
* this connector. */
encoder = drmModeGetEncoder(drm_fd, head->connector->encoder_id);
if (encoder != NULL) {
head->inherited_crtc_id = encoder->crtc_id;
crtc = drmModeGetCrtc(drm_fd, encoder->crtc_id);
drmModeFreeEncoder(encoder);
if (crtc == NULL)
return -1;
if (crtc->mode_valid)
head->inherited_mode = crtc->mode;
drmModeFreeCrtc(crtc);
}
return 0;
}
static int
drm_output_set_mode(struct weston_output *base,
enum weston_drm_backend_output_mode mode,
const char *modeline)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
struct drm_head *head = to_drm_head(weston_output_get_first_head(base));
struct drm_mode *current;
if (drm_output_update_modelist_from_heads(output) < 0)
return -1;
current = drm_output_choose_initial_mode(b, output, mode, modeline,
&head->inherited_mode);
if (!current)
return -1;
output->base.current_mode = &current->base;
output->base.current_mode->flags |= WL_OUTPUT_MODE_CURRENT;
/* Set native_ fields, so weston_output_mode_switch_to_native() works */
output->base.native_mode = output->base.current_mode;
output->base.native_scale = output->base.current_scale;
return 0;
}
static void
drm_output_set_gbm_format(struct weston_output *base,
const char *gbm_format)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
if (parse_gbm_format(gbm_format, b->gbm_format, &output->gbm_format) == -1)
output->gbm_format = b->gbm_format;
/* Without universal planes, we can't discover which formats are
* supported by the primary plane; we just hope that the GBM format
* works. */
if (!b->universal_planes)
output->scanout_plane->formats[0].format = output->gbm_format;
}
static void
drm_output_set_seat(struct weston_output *base,
const char *seat)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
setup_output_seat_constraint(b, &output->base,
seat ? seat : "");
}
static int
drm_output_init_gamma_size(struct drm_output *output)
{
struct drm_backend *backend = to_drm_backend(output->base.compositor);
drmModeCrtc *crtc;
assert(output->base.compositor);
assert(output->crtc_id != 0);
crtc = drmModeGetCrtc(backend->drm.fd, output->crtc_id);
if (!crtc)
return -1;
output->base.gamma_size = crtc->gamma_size;
drmModeFreeCrtc(crtc);
return 0;
}
static uint32_t
drm_head_get_possible_crtcs_mask(struct drm_head *head)
{
uint32_t possible_crtcs = 0;
drmModeEncoder *encoder;
int i;
for (i = 0; i < head->connector->count_encoders; i++) {
encoder = drmModeGetEncoder(head->backend->drm.fd,
head->connector->encoders[i]);
if (!encoder)
continue;
possible_crtcs |= encoder->possible_crtcs;
drmModeFreeEncoder(encoder);
}
return possible_crtcs;
}
static int
drm_crtc_get_index(drmModeRes *resources, uint32_t crtc_id)
{
int i;
for (i = 0; i < resources->count_crtcs; i++) {
if (resources->crtcs[i] == crtc_id)
return i;
}
assert(0 && "unknown crtc id");
return -1;
}
/** Pick a CRTC that might be able to drive all attached connectors
*
* @param output The output whose attached heads to include.
* @param resources The DRM KMS resources.
* @return CRTC index, or -1 on failure or not found.
*/
static int
drm_output_pick_crtc(struct drm_output *output, drmModeRes *resources)
{
struct drm_backend *backend;
struct weston_head *base;
struct drm_head *head;
uint32_t possible_crtcs = 0xffffffff;
int existing_crtc[32];
unsigned j, n = 0;
uint32_t crtc_id;
int best_crtc_index = -1;
int fallback_crtc_index = -1;
int i;
bool match;
backend = to_drm_backend(output->base.compositor);
/* This algorithm ignores drmModeEncoder::possible_clones restriction,
* because it is more often set wrong than not in the kernel. */
/* Accumulate a mask of possible crtcs and find existing routings. */
wl_list_for_each(base, &output->base.head_list, output_link) {
head = to_drm_head(base);
possible_crtcs &= drm_head_get_possible_crtcs_mask(head);
crtc_id = head->inherited_crtc_id;
if (crtc_id > 0 && n < ARRAY_LENGTH(existing_crtc))
existing_crtc[n++] = drm_crtc_get_index(resources,
crtc_id);
}
/* Find a crtc that could drive each connector individually at least,
* and prefer existing routings. */
for (i = 0; i < resources->count_crtcs; i++) {
crtc_id = resources->crtcs[i];
/* Could the crtc not drive each connector? */
if (!(possible_crtcs & (1 << i)))
continue;
/* Is the crtc already in use? */
if (drm_output_find_by_crtc(backend, crtc_id))
continue;
/* Try to preserve the existing CRTC -> connector routing;
* it makes initialisation faster, and also since we have a
* very dumb picking algorithm, may preserve a better
* choice. */
for (j = 0; j < n; j++) {
if (existing_crtc[j] == i)
return i;
}
/* Check if any other head had existing routing to this CRTC.
* If they did, this is not the best CRTC as it might be needed
* for another output we haven't enabled yet. */
match = false;
wl_list_for_each(base, &backend->compositor->head_list,
compositor_link) {
head = to_drm_head(base);
if (head->base.output == &output->base)
continue;
if (weston_head_is_enabled(&head->base))
continue;
if (head->inherited_crtc_id == crtc_id) {
match = true;
break;
}
}
if (!match)
best_crtc_index = i;
fallback_crtc_index = i;
}
if (best_crtc_index != -1)
return best_crtc_index;
if (fallback_crtc_index != -1)
return fallback_crtc_index;
/* Likely possible_crtcs was empty due to asking for clones,
* but since the DRM documentation says the kernel lies, let's
* pick one crtc anyway. Trial and error is the only way to
* be sure if something doesn't work. */
/* First pick any existing assignment. */
for (j = 0; j < n; j++) {
crtc_id = resources->crtcs[existing_crtc[j]];
if (!drm_output_find_by_crtc(backend, crtc_id))
return existing_crtc[j];
}
/* Otherwise pick any available crtc. */
for (i = 0; i < resources->count_crtcs; i++) {
crtc_id = resources->crtcs[i];
if (!drm_output_find_by_crtc(backend, crtc_id))
return i;
}
return -1;
}
/** Allocate a CRTC for the output
*
* @param output The output with no allocated CRTC.
* @param resources DRM KMS resources.
* @return 0 on success, -1 on failure.
*
* Finds a free CRTC that might drive the attached connectors, reserves the CRTC
* for the output, and loads the CRTC properties.
*
* Populates the cursor and scanout planes.
*
* On failure, the output remains without a CRTC.
*/
static int
drm_output_init_crtc(struct drm_output *output, drmModeRes *resources)
{
struct drm_backend *b = to_drm_backend(output->base.compositor);
drmModeObjectPropertiesPtr props;
int i;
assert(output->crtc_id == 0);
i = drm_output_pick_crtc(output, resources);
if (i < 0) {
weston_log("Output '%s': No available CRTCs.\n",
output->base.name);
return -1;
}
output->crtc_id = resources->crtcs[i];
output->pipe = i;
props = drmModeObjectGetProperties(b->drm.fd, output->crtc_id,
DRM_MODE_OBJECT_CRTC);
if (!props) {
weston_log("failed to get CRTC properties\n");
goto err_crtc;
}
drm_property_info_populate(b, crtc_props, output->props_crtc,
WDRM_CRTC__COUNT, props);
drmModeFreeObjectProperties(props);
output->scanout_plane =
drm_output_find_special_plane(b, output,
WDRM_PLANE_TYPE_PRIMARY);
if (!output->scanout_plane) {
weston_log("Failed to find primary plane for output %s\n",
output->base.name);
goto err_crtc;
}
/* Failing to find a cursor plane is not fatal, as we'll fall back
* to software cursor. */
output->cursor_plane =
drm_output_find_special_plane(b, output,
WDRM_PLANE_TYPE_CURSOR);
wl_array_remove_uint32(&b->unused_crtcs, output->crtc_id);
return 0;
err_crtc:
output->crtc_id = 0;
output->pipe = 0;
return -1;
}
/** Free the CRTC from the output
*
* @param output The output whose CRTC to deallocate.
*
* The CRTC reserved for the given output becomes free to use again.
*/
static void
drm_output_fini_crtc(struct drm_output *output)
{
struct drm_backend *b = to_drm_backend(output->base.compositor);
uint32_t *unused;
if (!b->universal_planes && !b->shutting_down) {
/* With universal planes, the 'special' planes are allocated at
* startup, freed at shutdown, and live on the plane list in
* between. We want the planes to continue to exist and be freed
* up for other outputs.
*
* Without universal planes, our special planes are
* pseudo-planes allocated at output creation, freed at output
* destruction, and not usable by other outputs.
*
* On the other hand, if the compositor is already shutting down,
* the plane has already been destroyed.
*/
if (output->cursor_plane)
drm_plane_destroy(output->cursor_plane);
if (output->scanout_plane)
drm_plane_destroy(output->scanout_plane);
}
drm_property_info_free(output->props_crtc, WDRM_CRTC__COUNT);
assert(output->crtc_id != 0);
unused = wl_array_add(&b->unused_crtcs, sizeof(*unused));
*unused = output->crtc_id;
/* Force resetting unused CRTCs */
b->state_invalid = true;
output->crtc_id = 0;
output->cursor_plane = NULL;
output->scanout_plane = NULL;
}
static void
drm_output_print_modes(struct drm_output *output)
{
struct weston_mode *m;
struct drm_mode *dm;
const char *aspect_ratio;
wl_list_for_each(m, &output->base.mode_list, link) {
dm = to_drm_mode(m);
aspect_ratio = aspect_ratio_to_string(m->aspect_ratio);
weston_log_continue(STAMP_SPACE "%dx%d@%.1f%s%s%s, %.1f MHz\n",
m->width, m->height, m->refresh / 1000.0,
aspect_ratio,
m->flags & WL_OUTPUT_MODE_PREFERRED ?
", preferred" : "",
m->flags & WL_OUTPUT_MODE_CURRENT ?
", current" : "",
dm->mode_info.clock / 1000.0);
}
}
static int
drm_output_enable(struct weston_output *base)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
drmModeRes *resources;
int ret;
resources = drmModeGetResources(b->drm.fd);
if (!resources) {
weston_log("drmModeGetResources failed\n");
return -1;
}
ret = drm_output_init_crtc(output, resources);
drmModeFreeResources(resources);
if (ret < 0)
return -1;
if (drm_output_init_gamma_size(output) < 0)
goto err;
if (b->pageflip_timeout)
drm_output_pageflip_timer_create(output);
if (b->use_pixman) {
if (drm_output_init_pixman(output, b) < 0) {
weston_log("Failed to init output pixman state\n");
goto err;
}
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_IMXG2D)
} else if (b->use_g2d) {
if (drm_output_init_g2d(output, b) < 0) {
weston_log("Failed to init output g2d state\n");
goto err;
}
#endif
#if defined(ENABLE_OPENGL)
} else if (drm_output_init_egl(output, b) < 0) {
weston_log("Failed to init output gl state\n");
goto err;
#endif
#endif
}
drm_output_init_backlight(output);
output->base.start_repaint_loop = drm_output_start_repaint_loop;
output->base.repaint = drm_output_repaint;
output->base.assign_planes = drm_assign_planes;
output->base.set_dpms = drm_set_dpms;
output->base.switch_mode = drm_output_switch_mode;
output->base.set_gamma = drm_output_set_gamma;
if (output->cursor_plane)
weston_compositor_stack_plane(b->compositor,
&output->cursor_plane->base,
NULL);
else
b->cursors_are_broken = 1;
weston_compositor_stack_plane(b->compositor,
&output->scanout_plane->base,
&b->compositor->primary_plane);
weston_log("Output %s (crtc %d) video modes:\n",
output->base.name, output->crtc_id);
drm_output_print_modes(output);
return 0;
err:
drm_output_fini_crtc(output);
return -1;
}
static void
drm_output_deinit(struct weston_output *base)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
if (b->use_pixman)
drm_output_fini_pixman(output);
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_IMXG2D)
else if (b->use_g2d)
drm_output_fini_g2d(output);
#endif
#if defined(ENABLE_OPENGL)
else
drm_output_fini_egl(output);
#endif
#endif
/* Since our planes are no longer in use anywhere, remove their base
* weston_plane's link from the plane stacking list, unless we're
* shutting down, in which case the plane has already been
* destroyed. */
if (!b->shutting_down) {
wl_list_remove(&output->scanout_plane->base.link);
wl_list_init(&output->scanout_plane->base.link);
if (output->cursor_plane) {
wl_list_remove(&output->cursor_plane->base.link);
wl_list_init(&output->cursor_plane->base.link);
/* Turn off hardware cursor */
drmModeSetCursor(b->drm.fd, output->crtc_id, 0, 0, 0);
}
}
drm_output_fini_crtc(output);
}
static void
drm_head_destroy(struct drm_head *head);
static void
drm_output_destroy(struct weston_output *base)
{
struct drm_output *output = to_drm_output(base);
struct drm_backend *b = to_drm_backend(base->compositor);
if (output->page_flip_pending || output->vblank_pending ||
output->atomic_complete_pending) {
output->destroy_pending = 1;
weston_log("destroy output while page flip pending\n");
return;
}
if (output->base.enabled)
drm_output_deinit(&output->base);
drm_mode_list_destroy(b, &output->base.mode_list);
if (output->pageflip_timer)
wl_event_source_remove(output->pageflip_timer);
weston_output_release(&output->base);
assert(!output->state_last);
drm_output_state_free(output->state_cur);
free(output);
}
static int
drm_output_disable(struct weston_output *base)
{
struct drm_output *output = to_drm_output(base);
if (output->page_flip_pending || output->vblank_pending ||
output->atomic_complete_pending) {
output->disable_pending = 1;
return -1;
}
weston_log("Disabling output %s\n", output->base.name);
if (output->base.enabled)
drm_output_deinit(&output->base);
output->disable_pending = 0;
return 0;
}
/**
* Update the list of unused connectors and CRTCs
*
* This keeps the unused_crtc arrays up to date.
*
* @param b Weston backend structure
* @param resources DRM resources for this device
*/
static void
drm_backend_update_unused_outputs(struct drm_backend *b, drmModeRes *resources)
{
int i;
wl_array_release(&b->unused_crtcs);
wl_array_init(&b->unused_crtcs);
for (i = 0; i < resources->count_crtcs; i++) {
struct drm_output *output;
uint32_t *crtc_id;
output = drm_output_find_by_crtc(b, resources->crtcs[i]);
if (output && output->base.enabled)
continue;
crtc_id = wl_array_add(&b->unused_crtcs, sizeof(*crtc_id));
*crtc_id = resources->crtcs[i];
}
}
/** Replace connector data and monitor information
*
* @param head The head to update.
* @param connector The connector data to be owned by the head, must match
* the head's connector ID.
* @return 0 on success, -1 on failure.
*
* Takes ownership of @c connector on success, not on failure.
*
* May schedule a heads changed call.
*/
static int
drm_head_assign_connector_info(struct drm_head *head,
drmModeConnector *connector)
{
drmModeObjectProperties *props;
const char *make = "unknown";
const char *model = "unknown";
const char *serial_number = "unknown";
assert(connector);
assert(head->connector_id == connector->connector_id);
props = drmModeObjectGetProperties(head->backend->drm.fd,
head->connector_id,
DRM_MODE_OBJECT_CONNECTOR);
if (!props) {
weston_log("Error: failed to get connector '%s' properties\n",
head->base.name);
return -1;
}
if (head->connector)
drmModeFreeConnector(head->connector);
head->connector = connector;
drm_property_info_populate(head->backend, connector_props,
head->props_conn,
WDRM_CONNECTOR__COUNT, props);
find_and_parse_output_edid(head, props, &make, &model, &serial_number);
weston_head_set_monitor_strings(&head->base, make, model, serial_number);
weston_head_set_subpixel(&head->base,
drm_subpixel_to_wayland(head->connector->subpixel));
weston_head_set_physical_size(&head->base, head->connector->mmWidth,
head->connector->mmHeight);
drmModeFreeObjectProperties(props);
/* Unknown connection status is assumed disconnected. */
weston_head_set_connection_status(&head->base,
head->connector->connection == DRM_MODE_CONNECTED);
return 0;
}
static void
drm_head_log_info(struct drm_head *head, const char *msg)
{
if (head->base.connected) {
weston_log("DRM: head '%s' %s, connector %d is connected, "
"EDID make '%s', model '%s', serial '%s'\n",
head->base.name, msg, head->connector_id,
head->base.make, head->base.model,
head->base.serial_number ?: "");
} else {
weston_log("DRM: head '%s' %s, connector %d is disconnected.\n",
head->base.name, msg, head->connector_id);
}
}
/** Update connector and monitor information
*
* @param head The head to update.
*
* Re-reads the DRM property lists for the connector and updates monitor
* information and connection status. This may schedule a heads changed call
* to the user.
*/
static void
drm_head_update_info(struct drm_head *head)
{
drmModeConnector *connector;
connector = drmModeGetConnector(head->backend->drm.fd,
head->connector_id);
if (!connector) {
weston_log("DRM: getting connector info for '%s' failed.\n",
head->base.name);
return;
}
if (drm_head_assign_connector_info(head, connector) < 0)
drmModeFreeConnector(connector);
if (head->base.device_changed)
drm_head_log_info(head, "updated");
}
/**
* Create a Weston head for a connector
*
* Given a DRM connector, create a matching drm_head structure and add it
* to Weston's head list.
*
* @param b Weston backend structure
* @param connector_id DRM connector ID for the head
* @param drm_device udev device pointer
* @returns The new head, or NULL on failure.
*/
static struct drm_head *
drm_head_create(struct drm_backend *backend, uint32_t connector_id,
struct udev_device *drm_device)
{
struct drm_head *head;
drmModeConnector *connector;
char *name;
head = zalloc(sizeof *head);
if (!head)
return NULL;
connector = drmModeGetConnector(backend->drm.fd, connector_id);
if (!connector)
goto err_alloc;
name = make_connector_name(connector);
if (!name)
goto err_alloc;
weston_head_init(&head->base, name);
free(name);
head->connector_id = connector_id;
head->backend = backend;
head->backlight = backlight_init(drm_device, connector->connector_type);
if (drm_head_assign_connector_info(head, connector) < 0)
goto err_init;
if (head->connector->connector_type == DRM_MODE_CONNECTOR_LVDS ||
head->connector->connector_type == DRM_MODE_CONNECTOR_eDP)
weston_head_set_internal(&head->base);
if (drm_head_read_current_setup(head, backend) < 0) {
weston_log("Failed to retrieve current mode from connector %d.\n",
head->connector_id);
/* Not fatal. */
}
weston_compositor_add_head(backend->compositor, &head->base);
drm_head_log_info(head, "found");
return head;
err_init:
weston_head_release(&head->base);
err_alloc:
if (connector)
drmModeFreeConnector(connector);
free(head);
return NULL;
}
static void
drm_head_destroy(struct drm_head *head)
{
weston_head_release(&head->base);
drm_property_info_free(head->props_conn, WDRM_CONNECTOR__COUNT);
drmModeFreeConnector(head->connector);
if (head->backlight)
backlight_destroy(head->backlight);
free(head);
}
/**
* Create a Weston output structure
*
* Create an "empty" drm_output. This is the implementation of
* weston_backend::create_output.
*
* Creating an output is usually followed by drm_output_attach_head()
* and drm_output_enable() to make use of it.
*
* @param compositor The compositor instance.
* @param name Name for the new output.
* @returns The output, or NULL on failure.
*/
static struct weston_output *
drm_output_create(struct weston_compositor *compositor, const char *name)
{
struct drm_backend *b = to_drm_backend(compositor);
struct drm_output *output;
output = zalloc(sizeof *output);
if (output == NULL)
return NULL;
weston_output_init(&output->base, compositor, name);
output->base.enable = drm_output_enable;
output->base.destroy = drm_output_destroy;
output->base.disable = drm_output_disable;
output->base.attach_head = drm_output_attach_head;
output->base.detach_head = drm_output_detach_head;
output->destroy_pending = 0;
output->disable_pending = 0;
output->state_cur = drm_output_state_alloc(output, NULL);
weston_compositor_add_pending_output(&output->base, b->compositor);
return &output->base;
}
static int
drm_backend_create_heads(struct drm_backend *b, struct udev_device *drm_device)
{
struct drm_head *head;
drmModeRes *resources;
int i;
resources = drmModeGetResources(b->drm.fd);
if (!resources) {
weston_log("drmModeGetResources failed\n");
return -1;
}
b->min_width = resources->min_width;
b->max_width = resources->max_width;
b->min_height = resources->min_height;
b->max_height = resources->max_height;
for (i = 0; i < resources->count_connectors; i++) {
uint32_t connector_id = resources->connectors[i];
head = drm_head_create(b, connector_id, drm_device);
if (!head) {
weston_log("DRM: failed to create head for connector %d.\n",
connector_id);
}
}
drm_backend_update_unused_outputs(b, resources);
drmModeFreeResources(resources);
return 0;
}
static void
drm_backend_update_heads(struct drm_backend *b, struct udev_device *drm_device)
{
drmModeRes *resources;
struct weston_head *base, *next;
struct drm_head *head;
int i;
resources = drmModeGetResources(b->drm.fd);
if (!resources) {
weston_log("drmModeGetResources failed\n");
return;
}
/* collect new connectors that have appeared, e.g. MST */
for (i = 0; i < resources->count_connectors; i++) {
uint32_t connector_id = resources->connectors[i];
head = drm_head_find_by_connector(b, connector_id);
if (head) {
drm_head_update_info(head);
} else {
head = drm_head_create(b, connector_id, drm_device);
if (!head)
weston_log("DRM: failed to create head for hot-added connector %d.\n",
connector_id);
}
}
/* Remove connectors that have disappeared. */
wl_list_for_each_safe(base, next,
&b->compositor->head_list, compositor_link) {
bool removed = true;
head = to_drm_head(base);
for (i = 0; i < resources->count_connectors; i++) {
if (resources->connectors[i] == head->connector_id) {
removed = false;
break;
}
}
if (!removed)
continue;
weston_log("DRM: head '%s' (connector %d) disappeared.\n",
head->base.name, head->connector_id);
drm_head_destroy(head);
}
drm_backend_update_unused_outputs(b, resources);
drmModeFreeResources(resources);
}
static int
udev_event_is_hotplug(struct drm_backend *b, struct udev_device *device)
{
const char *sysnum;
const char *val;
sysnum = udev_device_get_sysnum(device);
if (!sysnum || atoi(sysnum) != b->drm.id)
return 0;
val = udev_device_get_property_value(device, "HOTPLUG");
if (!val)
return 0;
return strcmp(val, "1") == 0;
}
static int
udev_drm_event(int fd, uint32_t mask, void *data)
{
struct drm_backend *b = data;
struct udev_device *event;
event = udev_monitor_receive_device(b->udev_monitor);
if (udev_event_is_hotplug(b, event))
drm_backend_update_heads(b, event);
udev_device_unref(event);
return 1;
}
static void
drm_destroy(struct weston_compositor *ec)
{
struct drm_backend *b = to_drm_backend(ec);
struct weston_head *base, *next;
udev_input_destroy(&b->input);
wl_event_source_remove(b->udev_drm_source);
wl_event_source_remove(b->drm_source);
b->shutting_down = true;
destroy_sprites(b);
weston_compositor_shutdown(ec);
wl_list_for_each_safe(base, next, &ec->head_list, compositor_link)
drm_head_destroy(to_drm_head(base));
if (b->gbm)
gbm_device_destroy(b->gbm);
udev_monitor_unref(b->udev_monitor);
udev_unref(b->udev);
weston_launcher_destroy(ec->launcher);
wl_array_release(&b->unused_crtcs);
close(b->drm.fd);
free(b->drm.filename);
free(b);
}
static void
session_notify(struct wl_listener *listener, void *data)
{
struct weston_compositor *compositor = data;
struct drm_backend *b = to_drm_backend(compositor);
struct drm_plane *plane;
struct drm_output *output;
if (compositor->session_active) {
weston_log("activating session\n");
weston_compositor_wake(compositor);
weston_compositor_damage_all(compositor);
b->state_invalid = true;
udev_input_enable(&b->input);
} else {
weston_log("deactivating session\n");
udev_input_disable(&b->input);
weston_compositor_offscreen(compositor);
/* If we have a repaint scheduled (either from a
* pending pageflip or the idle handler), make sure we
* cancel that so we don't try to pageflip when we're
* vt switched away. The OFFSCREEN state will prevent
* further attempts at repainting. When we switch
* back, we schedule a repaint, which will process
* pending frame callbacks. */
wl_list_for_each(output, &compositor->output_list, base.link) {
output->base.repaint_needed = false;
if (output->cursor_plane)
drmModeSetCursor(b->drm.fd, output->crtc_id,
0, 0, 0);
}
output = container_of(compositor->output_list.next,
struct drm_output, base.link);
wl_list_for_each(plane, &b->plane_list, link) {
if (plane->type != WDRM_PLANE_TYPE_OVERLAY)
continue;
drmModeSetPlane(b->drm.fd,
plane->plane_id,
output->crtc_id, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0);
}
}
}
/**
* Determines whether or not a device is capable of modesetting. If successful,
* sets b->drm.fd and b->drm.filename to the opened device.
*/
static bool
drm_device_is_kms(struct drm_backend *b, struct udev_device *device)
{
const char *filename = udev_device_get_devnode(device);
const char *sysnum = udev_device_get_sysnum(device);
drmModeRes *res;
int id, fd;
if (!filename)
return false;
fd = weston_launcher_open(b->compositor->launcher, filename, O_RDWR);
if (fd < 0)
return false;
res = drmModeGetResources(fd);
if (!res)
goto out_fd;
if (res->count_crtcs <= 0 || res->count_connectors <= 0 ||
res->count_encoders <= 0)
goto out_res;
if (sysnum)
id = atoi(sysnum);
if (!sysnum || id < 0) {
weston_log("couldn't get sysnum for device %s\n", filename);
goto out_res;
}
/* We can be called successfully on multiple devices; if we have,
* clean up old entries. */
if (b->drm.fd >= 0)
weston_launcher_close(b->compositor->launcher, b->drm.fd);
free(b->drm.filename);
b->drm.fd = fd;
b->drm.id = id;
b->drm.filename = strdup(filename);
drmModeFreeResources(res);
return true;
out_res:
drmModeFreeResources(res);
out_fd:
weston_launcher_close(b->compositor->launcher, fd);
return false;
}
/*
* Find primary GPU
* Some systems may have multiple DRM devices attached to a single seat. This
* function loops over all devices and tries to find a PCI device with the
* boot_vga sysfs attribute set to 1.
* If no such device is found, the first DRM device reported by udev is used.
* Devices are also vetted to make sure they are are capable of modesetting,
* rather than pure render nodes (GPU with no display), or pure
* memory-allocation devices (VGEM).
*/
static struct udev_device*
find_primary_gpu(struct drm_backend *b, const char *seat)
{
struct udev_enumerate *e;
struct udev_list_entry *entry;
const char *path, *device_seat, *id;
struct udev_device *device, *drm_device, *pci;
e = udev_enumerate_new(b->udev);
udev_enumerate_add_match_subsystem(e, "drm");
udev_enumerate_add_match_sysname(e, "card[0-9]*");
udev_enumerate_scan_devices(e);
drm_device = NULL;
udev_list_entry_foreach(entry, udev_enumerate_get_list_entry(e)) {
bool is_boot_vga = false;
path = udev_list_entry_get_name(entry);
device = udev_device_new_from_syspath(b->udev, path);
if (!device)
continue;
device_seat = udev_device_get_property_value(device, "ID_SEAT");
if (!device_seat)
device_seat = default_seat;
if (strcmp(device_seat, seat)) {
udev_device_unref(device);
continue;
}
pci = udev_device_get_parent_with_subsystem_devtype(device,
"pci", NULL);
if (pci) {
id = udev_device_get_sysattr_value(pci, "boot_vga");
if (id && !strcmp(id, "1"))
is_boot_vga = true;
}
/* If we already have a modesetting-capable device, and this
* device isn't our boot-VGA device, we aren't going to use
* it. */
if (!is_boot_vga && drm_device) {
udev_device_unref(device);
continue;
}
/* Make sure this device is actually capable of modesetting;
* if this call succeeds, b->drm.{fd,filename} will be set,
* and any old values freed. */
if (!drm_device_is_kms(b, device)) {
udev_device_unref(device);
continue;
}
/* There can only be one boot_vga device, and we try to use it
* at all costs. */
if (is_boot_vga) {
if (drm_device)
udev_device_unref(drm_device);
drm_device = device;
break;
}
/* Per the (!is_boot_vga && drm_device) test above, we only
* trump existing saved devices with boot-VGA devices, so if
* we end up here, this must be the first device we've seen. */
assert(!drm_device);
drm_device = device;
}
/* If we're returning a device to use, we must have an open FD for
* it. */
assert(!!drm_device == (b->drm.fd >= 0));
udev_enumerate_unref(e);
return drm_device;
}
static struct udev_device *
open_specific_drm_device(struct drm_backend *b, const char *name)
{
struct udev_device *device;
device = udev_device_new_from_subsystem_sysname(b->udev, "drm", name);
if (!device) {
weston_log("ERROR: could not open DRM device '%s'\n", name);
return NULL;
}
if (!drm_device_is_kms(b, device)) {
udev_device_unref(device);
weston_log("ERROR: DRM device '%s' is not a KMS device.\n", name);
return NULL;
}
/* If we're returning a device to use, we must have an open FD for
* it. */
assert(b->drm.fd >= 0);
return device;
}
static void
planes_binding(struct weston_keyboard *keyboard, const struct timespec *time,
uint32_t key, void *data)
{
struct drm_backend *b = data;
switch (key) {
case KEY_C:
b->cursors_are_broken ^= 1;
break;
case KEY_V:
b->sprites_are_broken ^= 1;
break;
case KEY_O:
b->sprites_hidden ^= 1;
break;
default:
break;
}
}
#ifdef BUILD_VAAPI_RECORDER
static void
recorder_destroy(struct drm_output *output)
{
vaapi_recorder_destroy(output->recorder);
output->recorder = NULL;
output->base.disable_planes--;
wl_list_remove(&output->recorder_frame_listener.link);
weston_log("[libva recorder] done\n");
}
static void
recorder_frame_notify(struct wl_listener *listener, void *data)
{
struct drm_output *output;
struct drm_backend *b;
int fd, ret;
output = container_of(listener, struct drm_output,
recorder_frame_listener);
b = to_drm_backend(output->base.compositor);
if (!output->recorder)
return;
ret = drmPrimeHandleToFD(b->drm.fd,
output->scanout_plane->state_cur->fb->handles[0],
DRM_CLOEXEC, &fd);
if (ret) {
weston_log("[libva recorder] "
"failed to create prime fd for front buffer\n");
return;
}
ret = vaapi_recorder_frame(output->recorder, fd,
output->scanout_plane->state_cur->fb->strides[0]);
if (ret < 0) {
weston_log("[libva recorder] aborted: %m\n");
recorder_destroy(output);
}
}
static void *
create_recorder(struct drm_backend *b, int width, int height,
const char *filename)
{
int fd;
drm_magic_t magic;
fd = open(b->drm.filename, O_RDWR | O_CLOEXEC);
if (fd < 0)
return NULL;
drmGetMagic(fd, &magic);
drmAuthMagic(b->drm.fd, magic);
return vaapi_recorder_create(fd, width, height, filename);
}
static void
recorder_binding(struct weston_keyboard *keyboard, const struct timespec *time,
uint32_t key, void *data)
{
struct drm_backend *b = data;
struct drm_output *output;
int width, height;
output = container_of(b->compositor->output_list.next,
struct drm_output, base.link);
if (!output->recorder) {
if (output->gbm_format != GBM_FORMAT_XRGB8888) {
weston_log("failed to start vaapi recorder: "
"output format not supported\n");
return;
}
width = output->base.current_mode->width;
height = output->base.current_mode->height;
output->recorder =
create_recorder(b, width, height, "capture.h264");
if (!output->recorder) {
weston_log("failed to create vaapi recorder\n");
return;
}
output->base.disable_planes++;
output->recorder_frame_listener.notify = recorder_frame_notify;
wl_signal_add(&output->base.frame_signal,
&output->recorder_frame_listener);
weston_output_schedule_repaint(&output->base);
weston_log("[libva recorder] initialized\n");
} else {
recorder_destroy(output);
}
}
#else
static void
recorder_binding(struct weston_keyboard *keyboard, const struct timespec *time,
uint32_t key, void *data)
{
weston_log("Compiled without libva support\n");
}
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
static void
switch_to_gl_renderer(struct drm_backend *b)
{
struct drm_output *output;
bool dmabuf_support_inited;
if (!b->use_pixman)
return;
dmabuf_support_inited = !!b->compositor->renderer->import_dmabuf;
weston_log("Switching to GL renderer\n");
b->gbm = create_gbm_device(b->drm.fd);
if (!b->gbm) {
weston_log("Failed to create gbm device. "
"Aborting renderer switch\n");
return;
}
wl_list_for_each(output, &b->compositor->output_list, base.link)
pixman_renderer_output_destroy(&output->base);
b->compositor->renderer->destroy(b->compositor);
if (drm_backend_create_gl_renderer(b) < 0) {
gbm_device_destroy(b->gbm);
weston_log("Failed to create GL renderer. Quitting.\n");
/* FIXME: we need a function to shutdown cleanly */
assert(0);
}
wl_list_for_each(output, &b->compositor->output_list, base.link)
drm_output_init_egl(output, b);
b->use_pixman = 0;
if (!dmabuf_support_inited && b->compositor->renderer->import_dmabuf) {
if (linux_dmabuf_setup(b->compositor) < 0)
weston_log("Error: initializing dmabuf "
"support failed.\n");
}
}
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
static void
renderer_switch_binding(struct weston_keyboard *keyboard,
const struct timespec *time, uint32_t key, void *data)
{
struct drm_backend *b =
to_drm_backend(keyboard->seat->compositor);
switch_to_gl_renderer(b);
}
#endif
static const struct weston_drm_output_api api = {
drm_output_set_mode,
drm_output_set_gbm_format,
drm_output_set_seat,
};
static void
hdr10_metadata_destroy(struct wl_client *client,
struct wl_resource *resource)
{
wl_resource_destroy(resource);
}
static void
hdr10_metadata_set_metadata(struct wl_client *client,
struct wl_resource *resource,
uint32_t eotf,
uint32_t type,
uint32_t display_primaries_red,
uint32_t display_primaries_green,
uint32_t display_primaries_blue,
uint32_t white_point,
uint32_t mastering_display_luminance,
uint32_t max_cll,
uint32_t max_fall)
{
struct weston_compositor *compositor = wl_resource_get_user_data(resource);
struct drm_backend *b = to_drm_backend(compositor);
struct hdr_static_metadata hdr_metadata;
hdr_metadata.eotf = eotf & 0xffff;
hdr_metadata.type = type & 0xffff;
hdr_metadata.display_primaries_x[0] = (display_primaries_red >> 16) & 0xffff;
hdr_metadata.display_primaries_y[0] = display_primaries_red & 0xffff;
hdr_metadata.display_primaries_x[1] = (display_primaries_green >> 16) & 0xffff;
hdr_metadata.display_primaries_y[1] = display_primaries_green & 0xffff;
hdr_metadata.display_primaries_x[2] = (display_primaries_blue >> 16) & 0xffff;
hdr_metadata.display_primaries_y[2] = display_primaries_blue & 0xffff;
hdr_metadata.white_point_x = (white_point >> 16) & 0xffff;
hdr_metadata.white_point_y = white_point & 0xffff;
hdr_metadata.max_mastering_display_luminance =
(mastering_display_luminance >> 16) & 0xffff;
hdr_metadata.min_mastering_display_luminance =
mastering_display_luminance & 0xffff;
hdr_metadata.max_cll = max_cll & 0xffff;
hdr_metadata.max_fall = max_fall & 0xffff;
drmModeCreatePropertyBlob(b->drm.fd, &hdr_metadata, sizeof(hdr_metadata), &b->hdr_blob_id);
}
static const struct zwp_hdr10_metadata_v1_interface hdr10_metadata_interface = {
hdr10_metadata_destroy,
hdr10_metadata_set_metadata,
};
static void
bind_hdr10_metadata(struct wl_client *client,
void *data, uint32_t version, uint32_t id)
{
struct wl_resource *resource;
struct weston_compositor *compositor = data;
resource = wl_resource_create(client, &zwp_hdr10_metadata_v1_interface,
version, id);
if (resource == NULL) {
wl_client_post_no_memory(client);
return;
}
wl_resource_set_implementation(resource, &hdr10_metadata_interface,
compositor, NULL);
}
static bool
drm_backend_is_hdr_supported(struct weston_compositor *compositor)
{
struct drm_output *output;
struct drm_head *head;
wl_list_for_each(output, &compositor->output_list, base.link) {
wl_list_for_each(head, &output->base.head_list, base.output_link) {
if (head->props_conn[WDRM_CONNECTOR_HDR10_METADATA].prop_id > 0)
return true;
}
}
wl_list_for_each(output, &compositor->pending_output_list, base.link) {
wl_list_for_each(head, &output->base.head_list, base.output_link) {
if (head->props_conn[WDRM_CONNECTOR_HDR10_METADATA].prop_id > 0)
return true;
}
}
return true;
}
static struct drm_backend *
drm_backend_create(struct weston_compositor *compositor,
struct weston_drm_backend_config *config)
{
struct drm_backend *b;
struct udev_device *drm_device;
struct wl_event_loop *loop;
const char *seat_id = default_seat;
const char *session_seat;
int ret;
session_seat = getenv("XDG_SEAT");
if (session_seat)
seat_id = session_seat;
if (config->seat_id)
seat_id = config->seat_id;
weston_log("initializing drm backend\n");
b = zalloc(sizeof *b);
if (b == NULL)
return NULL;
b->state_invalid = true;
b->drm.fd = -1;
wl_array_init(&b->unused_crtcs);
b->compositor = compositor;
b->use_pixman = config->use_pixman;
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
b->use_g2d = config->use_g2d;
#endif
b->pageflip_timeout = config->pageflip_timeout;
b->use_pixman_shadow = config->use_pixman_shadow;
compositor->backend = &b->base;
if (parse_gbm_format(config->gbm_format, GBM_FORMAT_XRGB8888, &b->gbm_format) < 0)
goto err_compositor;
/* Check if we run drm-backend using weston-launch */
compositor->launcher = weston_launcher_connect(compositor, config->tty,
seat_id, true);
if (compositor->launcher == NULL) {
weston_log("fatal: drm backend should be run using "
"weston-launch binary, or your system should "
"provide the logind D-Bus API.\n");
goto err_compositor;
}
b->udev = udev_new();
if (b->udev == NULL) {
weston_log("failed to initialize udev context\n");
goto err_launcher;
}
b->session_listener.notify = session_notify;
wl_signal_add(&compositor->session_signal, &b->session_listener);
if (config->specific_device)
drm_device = open_specific_drm_device(b, config->specific_device);
else
drm_device = find_primary_gpu(b, seat_id);
if (drm_device == NULL) {
weston_log("no drm device found\n");
goto err_udev;
}
if (init_kms_caps(b) < 0) {
weston_log("failed to initialize kms\n");
goto err_udev_dev;
}
if (b->use_pixman) {
if (init_pixman(b) < 0) {
weston_log("failed to initialize pixman renderer\n");
goto err_udev_dev;
}
#if defined(ENABLE_IMXGPU)
#if defined(ENABLE_IMXG2D)
} else if(b->use_g2d){
if (init_g2d(b) < 0) {
weston_log("failed to initialize g2d render\n");
goto err_udev_dev;
}
#endif
#if defined(ENABLE_OPENGL)
} else {
if (init_egl(b) < 0) {
weston_log("failed to initialize egl\n");
goto err_udev_dev;
}
#endif
#endif
}
b->base.destroy = drm_destroy;
b->base.repaint_begin = drm_repaint_begin;
b->base.repaint_flush = drm_repaint_flush;
b->base.repaint_cancel = drm_repaint_cancel;
b->base.create_output = drm_output_create;
b->base.query_dmabuf_formats = drm_query_dmabuf_formats;
b->base.import_dmabuf = drm_import_dmabuf;
weston_setup_vt_switch_bindings(compositor);
wl_list_init(&b->plane_list);
create_sprites(b);
if (udev_input_init(&b->input,
compositor, b->udev, seat_id,
config->configure_device) < 0) {
weston_log("failed to create input devices\n");
goto err_sprite;
}
if (drm_backend_create_heads(b, drm_device) < 0) {
weston_log("Failed to create heads for %s\n", b->drm.filename);
goto err_udev_input;
}
/* A this point we have some idea of whether or not we have a working
* cursor plane. */
if (!b->cursors_are_broken)
compositor->capabilities |= WESTON_CAP_CURSOR_PLANE;
loop = wl_display_get_event_loop(compositor->wl_display);
b->drm_source =
wl_event_loop_add_fd(loop, b->drm.fd,
WL_EVENT_READABLE, on_drm_input, b);
b->udev_monitor = udev_monitor_new_from_netlink(b->udev, "udev");
if (b->udev_monitor == NULL) {
weston_log("failed to initialize udev monitor\n");
goto err_drm_source;
}
udev_monitor_filter_add_match_subsystem_devtype(b->udev_monitor,
"drm", NULL);
b->udev_drm_source =
wl_event_loop_add_fd(loop,
udev_monitor_get_fd(b->udev_monitor),
WL_EVENT_READABLE, udev_drm_event, b);
if (udev_monitor_enable_receiving(b->udev_monitor) < 0) {
weston_log("failed to enable udev-monitor receiving\n");
goto err_udev_monitor;
}
udev_device_unref(drm_device);
weston_compositor_add_debug_binding(compositor, KEY_O,
planes_binding, b);
weston_compositor_add_debug_binding(compositor, KEY_C,
planes_binding, b);
weston_compositor_add_debug_binding(compositor, KEY_V,
planes_binding, b);
weston_compositor_add_debug_binding(compositor, KEY_Q,
recorder_binding, b);
#if defined(ENABLE_IMXGPU) && defined(ENABLE_OPENGL)
weston_compositor_add_debug_binding(compositor, KEY_W,
renderer_switch_binding, b);
#endif
if (compositor->renderer->import_dmabuf) {
if (linux_dmabuf_setup(compositor) < 0)
weston_log("Error: initializing dmabuf "
"support failed.\n");
}
if (drm_backend_is_hdr_supported(compositor)) {
if (!wl_global_create(compositor->wl_display, &zwp_hdr10_metadata_v1_interface, 1,
compositor, bind_hdr10_metadata)) {
weston_log("Error: initializing hdr10 support failed\n");
}
} else {
weston_log("info: HDR is not support\n");
}
ret = weston_plugin_api_register(compositor, WESTON_DRM_OUTPUT_API_NAME,
&api, sizeof(api));
if (ret < 0) {
weston_log("Failed to register output API.\n");
goto err_udev_monitor;
}
return b;
err_udev_monitor:
wl_event_source_remove(b->udev_drm_source);
udev_monitor_unref(b->udev_monitor);
err_drm_source:
wl_event_source_remove(b->drm_source);
err_udev_input:
udev_input_destroy(&b->input);
err_sprite:
if (b->gbm)
gbm_device_destroy(b->gbm);
destroy_sprites(b);
err_udev_dev:
udev_device_unref(drm_device);
err_launcher:
weston_launcher_destroy(compositor->launcher);
err_udev:
udev_unref(b->udev);
err_compositor:
weston_compositor_shutdown(compositor);
free(b);
return NULL;
}
static void
config_init_to_defaults(struct weston_drm_backend_config *config)
{
#if !defined(ENABLE_IMXGPU) || !defined(ENABLE_OPENGL) && !defined(ENABLE_IMXG2D)
config->use_pixman = 1;
config->use_pixman_shadow = true;
#else
config->use_pixman = 0;
config->use_pixman_shadow = false;
#endif
#if defined(ENABLE_IMXGPU) && defined(ENABLE_IMXG2D)
#if !defined(ENABLE_OPENGL)
config->use_g2d = 1;
#else
config->use_g2d = 0;
#endif
#endif
}
WL_EXPORT int
weston_backend_init(struct weston_compositor *compositor,
struct weston_backend_config *config_base)
{
struct drm_backend *b;
struct weston_drm_backend_config config = {{ 0, }};
if (config_base == NULL ||
config_base->struct_version != WESTON_DRM_BACKEND_CONFIG_VERSION ||
config_base->struct_size > sizeof(struct weston_drm_backend_config)) {
weston_log("drm backend config structure is invalid\n");
return -1;
}
config_init_to_defaults(&config);
memcpy(&config, config_base, config_base->struct_size);
b = drm_backend_create(compositor, &config);
if (b == NULL)
return -1;
return 0;
}