| /* |
| * Copyright © 2013 Keith Packard |
| * Copyright © 2015 Boyan Ding |
| * |
| * Permission to use, copy, modify, distribute, and sell this software and its |
| * documentation for any purpose is hereby granted without fee, provided that |
| * the above copyright notice appear in all copies and that both that copyright |
| * notice and this permission notice appear in supporting documentation, and |
| * that the name of the copyright holders not be used in advertising or |
| * publicity pertaining to distribution of the software without specific, |
| * written prior permission. The copyright holders make no representations |
| * about the suitability of this software for any purpose. It is provided "as |
| * is" without express or implied warranty. |
| * |
| * THE COPYRIGHT HOLDERS DISCLAIM ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, |
| * INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS, IN NO |
| * EVENT SHALL THE COPYRIGHT HOLDERS BE LIABLE FOR ANY SPECIAL, INDIRECT OR |
| * CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, |
| * DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER |
| * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE |
| * OF THIS SOFTWARE. |
| */ |
| |
| #include <fcntl.h> |
| #include <stdlib.h> |
| #include <unistd.h> |
| #include <string.h> |
| |
| #include <X11/xshmfence.h> |
| #include <xcb/xcb.h> |
| #include <xcb/dri3.h> |
| #include <xcb/present.h> |
| #include <xcb/xfixes.h> |
| |
| #include <X11/Xlib-xcb.h> |
| |
| #include "loader_dri_helper.h" |
| #include "loader_dri3_helper.h" |
| #include "util/macros.h" |
| #include "drm-uapi/drm_fourcc.h" |
| |
| /** |
| * A cached blit context. |
| */ |
| struct loader_dri3_blit_context { |
| mtx_t mtx; |
| __DRIcontext *ctx; |
| __DRIscreen *cur_screen; |
| const __DRIcoreExtension *core; |
| }; |
| |
| /* For simplicity we maintain the cache only for a single screen at a time */ |
| static struct loader_dri3_blit_context blit_context = { |
| _MTX_INITIALIZER_NP, NULL |
| }; |
| |
| static void |
| dri3_flush_present_events(struct loader_dri3_drawable *draw); |
| |
| static struct loader_dri3_buffer * |
| dri3_find_back_alloc(struct loader_dri3_drawable *draw); |
| |
| static xcb_screen_t * |
| get_screen_for_root(xcb_connection_t *conn, xcb_window_t root) |
| { |
| xcb_screen_iterator_t screen_iter = |
| xcb_setup_roots_iterator(xcb_get_setup(conn)); |
| |
| for (; screen_iter.rem; xcb_screen_next (&screen_iter)) { |
| if (screen_iter.data->root == root) |
| return screen_iter.data; |
| } |
| |
| return NULL; |
| } |
| |
| static xcb_visualtype_t * |
| get_xcb_visualtype_for_depth(struct loader_dri3_drawable *draw, int depth) |
| { |
| xcb_visualtype_iterator_t visual_iter; |
| xcb_screen_t *screen = draw->screen; |
| xcb_depth_iterator_t depth_iter; |
| |
| if (!screen) |
| return NULL; |
| |
| depth_iter = xcb_screen_allowed_depths_iterator(screen); |
| for (; depth_iter.rem; xcb_depth_next(&depth_iter)) { |
| if (depth_iter.data->depth != depth) |
| continue; |
| |
| visual_iter = xcb_depth_visuals_iterator(depth_iter.data); |
| if (visual_iter.rem) |
| return visual_iter.data; |
| } |
| |
| return NULL; |
| } |
| |
| /* Sets the adaptive sync window property state. */ |
| static void |
| set_adaptive_sync_property(xcb_connection_t *conn, xcb_drawable_t drawable, |
| uint32_t state) |
| { |
| static char const name[] = "_VARIABLE_REFRESH"; |
| xcb_intern_atom_cookie_t cookie; |
| xcb_intern_atom_reply_t* reply; |
| xcb_void_cookie_t check; |
| |
| cookie = xcb_intern_atom(conn, 0, strlen(name), name); |
| reply = xcb_intern_atom_reply(conn, cookie, NULL); |
| if (reply == NULL) |
| return; |
| |
| if (state) |
| check = xcb_change_property_checked(conn, XCB_PROP_MODE_REPLACE, |
| drawable, reply->atom, |
| XCB_ATOM_CARDINAL, 32, 1, &state); |
| else |
| check = xcb_delete_property_checked(conn, drawable, reply->atom); |
| |
| xcb_discard_reply(conn, check.sequence); |
| free(reply); |
| } |
| |
| /* Get red channel mask for given drawable at given depth. */ |
| static unsigned int |
| dri3_get_red_mask_for_depth(struct loader_dri3_drawable *draw, int depth) |
| { |
| xcb_visualtype_t *visual = get_xcb_visualtype_for_depth(draw, depth); |
| |
| if (visual) |
| return visual->red_mask; |
| |
| return 0; |
| } |
| |
| /** |
| * Do we have blit functionality in the image blit extension? |
| * |
| * \param draw[in] The drawable intended to blit from / to. |
| * \return true if we have blit functionality. false otherwise. |
| */ |
| static bool loader_dri3_have_image_blit(const struct loader_dri3_drawable *draw) |
| { |
| return draw->ext->image->base.version >= 9 && |
| draw->ext->image->blitImage != NULL; |
| } |
| |
| /** |
| * Get and lock (for use with the current thread) a dri context associated |
| * with the drawable's dri screen. The context is intended to be used with |
| * the dri image extension's blitImage method. |
| * |
| * \param draw[in] Pointer to the drawable whose dri screen we want a |
| * dri context for. |
| * \return A dri context or NULL if context creation failed. |
| * |
| * When the caller is done with the context (even if the context returned was |
| * NULL), the caller must call loader_dri3_blit_context_put. |
| */ |
| static __DRIcontext * |
| loader_dri3_blit_context_get(struct loader_dri3_drawable *draw) |
| { |
| mtx_lock(&blit_context.mtx); |
| |
| if (blit_context.ctx && blit_context.cur_screen != draw->dri_screen) { |
| blit_context.core->destroyContext(blit_context.ctx); |
| blit_context.ctx = NULL; |
| } |
| |
| if (!blit_context.ctx) { |
| blit_context.ctx = draw->ext->core->createNewContext(draw->dri_screen, |
| NULL, NULL, NULL); |
| blit_context.cur_screen = draw->dri_screen; |
| blit_context.core = draw->ext->core; |
| } |
| |
| return blit_context.ctx; |
| } |
| |
| /** |
| * Release (for use with other threads) a dri context previously obtained using |
| * loader_dri3_blit_context_get. |
| */ |
| static void |
| loader_dri3_blit_context_put(void) |
| { |
| mtx_unlock(&blit_context.mtx); |
| } |
| |
| /** |
| * Blit (parts of) the contents of a DRI image to another dri image |
| * |
| * \param draw[in] The drawable which owns the images. |
| * \param dst[in] The destination image. |
| * \param src[in] The source image. |
| * \param dstx0[in] Start destination coordinate. |
| * \param dsty0[in] Start destination coordinate. |
| * \param width[in] Blit width. |
| * \param height[in] Blit height. |
| * \param srcx0[in] Start source coordinate. |
| * \param srcy0[in] Start source coordinate. |
| * \param flush_flag[in] Image blit flush flag. |
| * \return true iff successful. |
| */ |
| static bool |
| loader_dri3_blit_image(struct loader_dri3_drawable *draw, |
| __DRIimage *dst, __DRIimage *src, |
| int dstx0, int dsty0, int width, int height, |
| int srcx0, int srcy0, int flush_flag) |
| { |
| __DRIcontext *dri_context; |
| bool use_blit_context = false; |
| |
| if (!loader_dri3_have_image_blit(draw)) |
| return false; |
| |
| dri_context = draw->vtable->get_dri_context(draw); |
| |
| if (!dri_context || !draw->vtable->in_current_context(draw)) { |
| dri_context = loader_dri3_blit_context_get(draw); |
| use_blit_context = true; |
| flush_flag |= __BLIT_FLAG_FLUSH; |
| } |
| |
| if (dri_context) |
| draw->ext->image->blitImage(dri_context, dst, src, dstx0, dsty0, |
| width, height, srcx0, srcy0, |
| width, height, flush_flag); |
| |
| if (use_blit_context) |
| loader_dri3_blit_context_put(); |
| |
| return dri_context != NULL; |
| } |
| |
| static inline void |
| dri3_fence_reset(xcb_connection_t *c, struct loader_dri3_buffer *buffer) |
| { |
| xshmfence_reset(buffer->shm_fence); |
| } |
| |
| static inline void |
| dri3_fence_set(struct loader_dri3_buffer *buffer) |
| { |
| xshmfence_trigger(buffer->shm_fence); |
| } |
| |
| static inline void |
| dri3_fence_trigger(xcb_connection_t *c, struct loader_dri3_buffer *buffer) |
| { |
| xcb_sync_trigger_fence(c, buffer->sync_fence); |
| } |
| |
| static inline void |
| dri3_fence_await(xcb_connection_t *c, struct loader_dri3_drawable *draw, |
| struct loader_dri3_buffer *buffer) |
| { |
| xcb_flush(c); |
| xshmfence_await(buffer->shm_fence); |
| if (draw) { |
| mtx_lock(&draw->mtx); |
| dri3_flush_present_events(draw); |
| mtx_unlock(&draw->mtx); |
| } |
| } |
| |
| static void |
| dri3_update_max_num_back(struct loader_dri3_drawable *draw) |
| { |
| switch (draw->last_present_mode) { |
| case XCB_PRESENT_COMPLETE_MODE_FLIP: { |
| int new_max; |
| |
| if (draw->swap_interval == 0) |
| new_max = 4; |
| else |
| new_max = 3; |
| |
| assert(new_max <= LOADER_DRI3_MAX_BACK); |
| |
| if (new_max != draw->max_num_back) { |
| /* On transition from swap interval == 0 to != 0, start with two |
| * buffers again. Otherwise keep the current number of buffers. Either |
| * way, more will be allocated if needed. |
| */ |
| if (new_max < draw->max_num_back) |
| draw->cur_num_back = 2; |
| |
| draw->max_num_back = new_max; |
| } |
| |
| break; |
| } |
| |
| case XCB_PRESENT_COMPLETE_MODE_SKIP: |
| break; |
| |
| default: |
| /* On transition from flips to copies, start with a single buffer again, |
| * a second one will be allocated if needed |
| */ |
| if (draw->max_num_back != 2) |
| draw->cur_num_back = 1; |
| |
| draw->max_num_back = 2; |
| } |
| } |
| |
| void |
| loader_dri3_set_swap_interval(struct loader_dri3_drawable *draw, int interval) |
| { |
| /* Wait all previous swap done before changing swap interval. |
| * |
| * This is for preventing swap out of order in the following cases: |
| * 1. Change from sync swap mode (>0) to async mode (=0), so async swap occurs |
| * before previous pending sync swap. |
| * 2. Change from value A to B and A > B, so the target_msc for the previous |
| * pending swap may be bigger than newer swap. |
| * |
| * PS. changing from value A to B and A < B won't cause swap out of order but |
| * may still gets wrong target_msc value at the beginning. |
| */ |
| if (draw->swap_interval != interval) |
| loader_dri3_swapbuffer_barrier(draw); |
| |
| draw->swap_interval = interval; |
| } |
| |
| /** dri3_free_render_buffer |
| * |
| * Free everything associated with one render buffer including pixmap, fence |
| * stuff and the driver image |
| */ |
| static void |
| dri3_free_render_buffer(struct loader_dri3_drawable *draw, |
| struct loader_dri3_buffer *buffer) |
| { |
| if (buffer->own_pixmap) |
| xcb_free_pixmap(draw->conn, buffer->pixmap); |
| xcb_sync_destroy_fence(draw->conn, buffer->sync_fence); |
| xshmfence_unmap_shm(buffer->shm_fence); |
| draw->ext->image->destroyImage(buffer->image); |
| if (buffer->linear_buffer) |
| draw->ext->image->destroyImage(buffer->linear_buffer); |
| free(buffer); |
| } |
| |
| void |
| loader_dri3_drawable_fini(struct loader_dri3_drawable *draw) |
| { |
| int i; |
| |
| draw->ext->core->destroyDrawable(draw->dri_drawable); |
| |
| for (i = 0; i < ARRAY_SIZE(draw->buffers); i++) { |
| if (draw->buffers[i]) |
| dri3_free_render_buffer(draw, draw->buffers[i]); |
| } |
| |
| if (draw->special_event) { |
| xcb_void_cookie_t cookie = |
| xcb_present_select_input_checked(draw->conn, draw->eid, draw->drawable, |
| XCB_PRESENT_EVENT_MASK_NO_EVENT); |
| |
| xcb_discard_reply(draw->conn, cookie.sequence); |
| xcb_unregister_for_special_event(draw->conn, draw->special_event); |
| } |
| |
| if (draw->region) |
| xcb_xfixes_destroy_region(draw->conn, draw->region); |
| |
| cnd_destroy(&draw->event_cnd); |
| mtx_destroy(&draw->mtx); |
| } |
| |
| int |
| loader_dri3_drawable_init(xcb_connection_t *conn, |
| xcb_drawable_t drawable, |
| enum loader_dri3_drawable_type type, |
| __DRIscreen *dri_screen, |
| bool is_different_gpu, |
| bool multiplanes_available, |
| bool prefer_back_buffer_reuse, |
| const __DRIconfig *dri_config, |
| struct loader_dri3_extensions *ext, |
| const struct loader_dri3_vtable *vtable, |
| struct loader_dri3_drawable *draw) |
| { |
| xcb_get_geometry_cookie_t cookie; |
| xcb_get_geometry_reply_t *reply; |
| xcb_generic_error_t *error; |
| |
| draw->conn = conn; |
| draw->ext = ext; |
| draw->vtable = vtable; |
| draw->drawable = drawable; |
| draw->type = type; |
| draw->region = 0; |
| draw->dri_screen = dri_screen; |
| draw->is_different_gpu = is_different_gpu; |
| draw->multiplanes_available = multiplanes_available; |
| draw->prefer_back_buffer_reuse = prefer_back_buffer_reuse; |
| draw->queries_buffer_age = false; |
| |
| draw->have_back = 0; |
| draw->have_fake_front = 0; |
| draw->first_init = true; |
| draw->adaptive_sync = false; |
| draw->adaptive_sync_active = false; |
| draw->block_on_depleted_buffers = false; |
| |
| draw->cur_blit_source = -1; |
| draw->back_format = __DRI_IMAGE_FORMAT_NONE; |
| mtx_init(&draw->mtx, mtx_plain); |
| cnd_init(&draw->event_cnd); |
| |
| if (draw->ext->config) { |
| unsigned char adaptive_sync = 0; |
| unsigned char block_on_depleted_buffers = 0; |
| |
| draw->ext->config->configQueryb(draw->dri_screen, |
| "adaptive_sync", |
| &adaptive_sync); |
| |
| draw->adaptive_sync = adaptive_sync; |
| |
| draw->ext->config->configQueryb(draw->dri_screen, |
| "block_on_depleted_buffers", |
| &block_on_depleted_buffers); |
| |
| draw->block_on_depleted_buffers = block_on_depleted_buffers; |
| } |
| |
| if (!draw->adaptive_sync) |
| set_adaptive_sync_property(conn, draw->drawable, false); |
| |
| draw->swap_interval = dri_get_initial_swap_interval(draw->dri_screen, |
| draw->ext->config); |
| |
| dri3_update_max_num_back(draw); |
| |
| /* Create a new drawable */ |
| draw->dri_drawable = |
| draw->ext->image_driver->createNewDrawable(dri_screen, |
| dri_config, |
| draw); |
| |
| if (!draw->dri_drawable) |
| return 1; |
| |
| cookie = xcb_get_geometry(draw->conn, draw->drawable); |
| reply = xcb_get_geometry_reply(draw->conn, cookie, &error); |
| if (reply == NULL || error != NULL) { |
| draw->ext->core->destroyDrawable(draw->dri_drawable); |
| return 1; |
| } |
| |
| draw->screen = get_screen_for_root(draw->conn, reply->root); |
| draw->width = reply->width; |
| draw->height = reply->height; |
| draw->depth = reply->depth; |
| draw->vtable->set_drawable_size(draw, draw->width, draw->height); |
| free(reply); |
| |
| draw->swap_method = __DRI_ATTRIB_SWAP_UNDEFINED; |
| if (draw->ext->core->base.version >= 2) { |
| (void )draw->ext->core->getConfigAttrib(dri_config, |
| __DRI_ATTRIB_SWAP_METHOD, |
| &draw->swap_method); |
| } |
| |
| /* |
| * Make sure server has the same swap interval we do for the new |
| * drawable. |
| */ |
| loader_dri3_set_swap_interval(draw, draw->swap_interval); |
| |
| return 0; |
| } |
| |
| /* XXX this belongs in presentproto */ |
| #ifndef PresentWindowDestroyed |
| #define PresentWindowDestroyed (1 << 0) |
| #endif |
| /* |
| * Process one Present event |
| */ |
| static bool |
| dri3_handle_present_event(struct loader_dri3_drawable *draw, |
| xcb_present_generic_event_t *ge) |
| { |
| switch (ge->evtype) { |
| case XCB_PRESENT_CONFIGURE_NOTIFY: { |
| xcb_present_configure_notify_event_t *ce = (void *) ge; |
| if (ce->pixmap_flags & PresentWindowDestroyed) { |
| free(ge); |
| return false; |
| } |
| |
| draw->width = ce->width; |
| draw->height = ce->height; |
| draw->vtable->set_drawable_size(draw, draw->width, draw->height); |
| draw->ext->flush->invalidate(draw->dri_drawable); |
| break; |
| } |
| case XCB_PRESENT_COMPLETE_NOTIFY: { |
| xcb_present_complete_notify_event_t *ce = (void *) ge; |
| |
| /* Compute the processed SBC number from the received 32-bit serial number |
| * merged with the upper 32-bits of the sent 64-bit serial number while |
| * checking for wrap. |
| */ |
| if (ce->kind == XCB_PRESENT_COMPLETE_KIND_PIXMAP) { |
| uint64_t recv_sbc = (draw->send_sbc & 0xffffffff00000000LL) | ce->serial; |
| |
| /* Only assume wraparound if that results in exactly the previous |
| * SBC + 1, otherwise ignore received SBC > sent SBC (those are |
| * probably from a previous loader_dri3_drawable instance) to avoid |
| * calculating bogus target MSC values in loader_dri3_swap_buffers_msc |
| */ |
| if (recv_sbc <= draw->send_sbc) |
| draw->recv_sbc = recv_sbc; |
| else if (recv_sbc == (draw->recv_sbc + 0x100000001ULL)) |
| draw->recv_sbc = recv_sbc - 0x100000000ULL; |
| |
| /* When moving from flip to copy, we assume that we can allocate in |
| * a more optimal way if we don't need to cater for the display |
| * controller. |
| */ |
| if (ce->mode == XCB_PRESENT_COMPLETE_MODE_COPY && |
| draw->last_present_mode == XCB_PRESENT_COMPLETE_MODE_FLIP) { |
| for (int b = 0; b < ARRAY_SIZE(draw->buffers); b++) { |
| if (draw->buffers[b]) |
| draw->buffers[b]->reallocate = true; |
| } |
| } |
| |
| /* If the server tells us that our allocation is suboptimal, we |
| * reallocate once. |
| */ |
| #ifdef HAVE_DRI3_MODIFIERS |
| if (ce->mode == XCB_PRESENT_COMPLETE_MODE_SUBOPTIMAL_COPY && |
| draw->last_present_mode != ce->mode) { |
| for (int b = 0; b < ARRAY_SIZE(draw->buffers); b++) { |
| if (draw->buffers[b]) |
| draw->buffers[b]->reallocate = true; |
| } |
| } |
| #endif |
| draw->last_present_mode = ce->mode; |
| |
| if (draw->vtable->show_fps) |
| draw->vtable->show_fps(draw, ce->ust); |
| |
| draw->ust = ce->ust; |
| draw->msc = ce->msc; |
| } else if (ce->serial == draw->eid) { |
| draw->notify_ust = ce->ust; |
| draw->notify_msc = ce->msc; |
| } |
| break; |
| } |
| case XCB_PRESENT_EVENT_IDLE_NOTIFY: { |
| xcb_present_idle_notify_event_t *ie = (void *) ge; |
| int b; |
| |
| for (b = 0; b < ARRAY_SIZE(draw->buffers); b++) { |
| struct loader_dri3_buffer *buf = draw->buffers[b]; |
| |
| if (buf && buf->pixmap == ie->pixmap) |
| buf->busy = 0; |
| } |
| break; |
| } |
| } |
| free(ge); |
| return true; |
| } |
| |
| static bool |
| dri3_wait_for_event_locked(struct loader_dri3_drawable *draw, |
| unsigned *full_sequence) |
| { |
| xcb_generic_event_t *ev; |
| xcb_present_generic_event_t *ge; |
| |
| xcb_flush(draw->conn); |
| |
| /* Only have one thread waiting for events at a time */ |
| if (draw->has_event_waiter) { |
| cnd_wait(&draw->event_cnd, &draw->mtx); |
| if (full_sequence) |
| *full_sequence = draw->last_special_event_sequence; |
| /* Another thread has updated the protected info, so retest. */ |
| return true; |
| } else { |
| draw->has_event_waiter = true; |
| /* Allow other threads access to the drawable while we're waiting. */ |
| mtx_unlock(&draw->mtx); |
| ev = xcb_wait_for_special_event(draw->conn, draw->special_event); |
| mtx_lock(&draw->mtx); |
| draw->has_event_waiter = false; |
| cnd_broadcast(&draw->event_cnd); |
| } |
| if (!ev) |
| return false; |
| draw->last_special_event_sequence = ev->full_sequence; |
| if (full_sequence) |
| *full_sequence = ev->full_sequence; |
| ge = (void *) ev; |
| return dri3_handle_present_event(draw, ge); |
| } |
| |
| /** loader_dri3_wait_for_msc |
| * |
| * Get the X server to send an event when the target msc/divisor/remainder is |
| * reached. |
| */ |
| bool |
| loader_dri3_wait_for_msc(struct loader_dri3_drawable *draw, |
| int64_t target_msc, |
| int64_t divisor, int64_t remainder, |
| int64_t *ust, int64_t *msc, int64_t *sbc) |
| { |
| xcb_void_cookie_t cookie = xcb_present_notify_msc(draw->conn, |
| draw->drawable, |
| draw->eid, |
| target_msc, |
| divisor, |
| remainder); |
| unsigned full_sequence; |
| |
| mtx_lock(&draw->mtx); |
| |
| /* Wait for the event */ |
| do { |
| if (!dri3_wait_for_event_locked(draw, &full_sequence)) { |
| mtx_unlock(&draw->mtx); |
| return false; |
| } |
| } while (full_sequence != cookie.sequence || draw->notify_msc < target_msc); |
| |
| *ust = draw->notify_ust; |
| *msc = draw->notify_msc; |
| *sbc = draw->recv_sbc; |
| mtx_unlock(&draw->mtx); |
| |
| return true; |
| } |
| |
| /** loader_dri3_wait_for_sbc |
| * |
| * Wait for the completed swap buffer count to reach the specified |
| * target. Presumably the application knows that this will be reached with |
| * outstanding complete events, or we're going to be here awhile. |
| */ |
| int |
| loader_dri3_wait_for_sbc(struct loader_dri3_drawable *draw, |
| int64_t target_sbc, int64_t *ust, |
| int64_t *msc, int64_t *sbc) |
| { |
| /* From the GLX_OML_sync_control spec: |
| * |
| * "If <target_sbc> = 0, the function will block until all previous |
| * swaps requested with glXSwapBuffersMscOML for that window have |
| * completed." |
| */ |
| mtx_lock(&draw->mtx); |
| if (!target_sbc) |
| target_sbc = draw->send_sbc; |
| |
| while (draw->recv_sbc < target_sbc) { |
| if (!dri3_wait_for_event_locked(draw, NULL)) { |
| mtx_unlock(&draw->mtx); |
| return 0; |
| } |
| } |
| |
| *ust = draw->ust; |
| *msc = draw->msc; |
| *sbc = draw->recv_sbc; |
| mtx_unlock(&draw->mtx); |
| return 1; |
| } |
| |
| /** loader_dri3_find_back |
| * |
| * Find an idle back buffer. If there isn't one, then |
| * wait for a present idle notify event from the X server |
| */ |
| static int |
| dri3_find_back(struct loader_dri3_drawable *draw, bool prefer_a_different) |
| { |
| int b; |
| int num_to_consider; |
| int max_num; |
| |
| mtx_lock(&draw->mtx); |
| /* Increase the likelyhood of reusing current buffer */ |
| dri3_flush_present_events(draw); |
| |
| /* Check whether we need to reuse the current back buffer as new back. |
| * In that case, wait until it's not busy anymore. |
| */ |
| if (!loader_dri3_have_image_blit(draw) && draw->cur_blit_source != -1) { |
| num_to_consider = 1; |
| max_num = 1; |
| draw->cur_blit_source = -1; |
| } else { |
| num_to_consider = draw->cur_num_back; |
| max_num = draw->max_num_back; |
| } |
| |
| /* In a DRI_PRIME situation, if prefer_a_different is true, we first try |
| * to find an idle buffer that is not the last used one. |
| * This is useful if we receive a XCB_PRESENT_EVENT_IDLE_NOTIFY event |
| * for a pixmap but it's not actually idle (eg: the DRI_PRIME blit is |
| * still in progress). |
| * Unigine Superposition hits this and this allows to use 2 back buffers |
| * instead of reusing the same one all the time, causing the next frame |
| * to wait for the copy to finish. |
| */ |
| int current_back_id = draw->cur_back; |
| for (;;) { |
| for (b = 0; b < num_to_consider; b++) { |
| int id = LOADER_DRI3_BACK_ID((b + draw->cur_back) % draw->cur_num_back); |
| struct loader_dri3_buffer *buffer = draw->buffers[id]; |
| |
| if (!buffer || (!buffer->busy && |
| (!prefer_a_different || id != current_back_id))) { |
| draw->cur_back = id; |
| mtx_unlock(&draw->mtx); |
| return id; |
| } |
| } |
| |
| if (num_to_consider < max_num) { |
| num_to_consider = ++draw->cur_num_back; |
| } else if (prefer_a_different) { |
| prefer_a_different = false; |
| } else if (!dri3_wait_for_event_locked(draw, NULL)) { |
| mtx_unlock(&draw->mtx); |
| return -1; |
| } |
| } |
| } |
| |
| static xcb_gcontext_t |
| dri3_drawable_gc(struct loader_dri3_drawable *draw) |
| { |
| if (!draw->gc) { |
| uint32_t v = 0; |
| xcb_create_gc(draw->conn, |
| (draw->gc = xcb_generate_id(draw->conn)), |
| draw->drawable, |
| XCB_GC_GRAPHICS_EXPOSURES, |
| &v); |
| } |
| return draw->gc; |
| } |
| |
| |
| static struct loader_dri3_buffer * |
| dri3_back_buffer(struct loader_dri3_drawable *draw) |
| { |
| return draw->buffers[LOADER_DRI3_BACK_ID(draw->cur_back)]; |
| } |
| |
| static struct loader_dri3_buffer * |
| dri3_front_buffer(struct loader_dri3_drawable *draw) |
| { |
| return draw->buffers[LOADER_DRI3_FRONT_ID]; |
| } |
| |
| static void |
| dri3_copy_area(xcb_connection_t *c, |
| xcb_drawable_t src_drawable, |
| xcb_drawable_t dst_drawable, |
| xcb_gcontext_t gc, |
| int16_t src_x, |
| int16_t src_y, |
| int16_t dst_x, |
| int16_t dst_y, |
| uint16_t width, |
| uint16_t height) |
| { |
| xcb_void_cookie_t cookie; |
| |
| cookie = xcb_copy_area_checked(c, |
| src_drawable, |
| dst_drawable, |
| gc, |
| src_x, |
| src_y, |
| dst_x, |
| dst_y, |
| width, |
| height); |
| xcb_discard_reply(c, cookie.sequence); |
| } |
| |
| /** |
| * Asks the driver to flush any queued work necessary for serializing with the |
| * X command stream, and optionally the slightly more strict requirement of |
| * glFlush() equivalence (which would require flushing even if nothing had |
| * been drawn to a window system framebuffer, for example). |
| */ |
| void |
| loader_dri3_flush(struct loader_dri3_drawable *draw, |
| unsigned flags, |
| enum __DRI2throttleReason throttle_reason) |
| { |
| /* NEED TO CHECK WHETHER CONTEXT IS NULL */ |
| __DRIcontext *dri_context = draw->vtable->get_dri_context(draw); |
| |
| if (dri_context) { |
| draw->ext->flush->flush_with_flags(dri_context, draw->dri_drawable, |
| flags, throttle_reason); |
| } |
| } |
| |
| void |
| loader_dri3_copy_sub_buffer(struct loader_dri3_drawable *draw, |
| int x, int y, |
| int width, int height, |
| bool flush) |
| { |
| struct loader_dri3_buffer *back; |
| unsigned flags = __DRI2_FLUSH_DRAWABLE; |
| |
| /* Check we have the right attachments */ |
| if (!draw->have_back || draw->type != LOADER_DRI3_DRAWABLE_WINDOW) |
| return; |
| |
| if (flush) |
| flags |= __DRI2_FLUSH_CONTEXT; |
| loader_dri3_flush(draw, flags, __DRI2_THROTTLE_COPYSUBBUFFER); |
| |
| back = dri3_find_back_alloc(draw); |
| if (!back) |
| return; |
| |
| y = draw->height - y - height; |
| |
| if (draw->is_different_gpu) { |
| /* Update the linear buffer part of the back buffer |
| * for the dri3_copy_area operation |
| */ |
| (void) loader_dri3_blit_image(draw, |
| back->linear_buffer, |
| back->image, |
| 0, 0, back->width, back->height, |
| 0, 0, __BLIT_FLAG_FLUSH); |
| } |
| |
| loader_dri3_swapbuffer_barrier(draw); |
| dri3_fence_reset(draw->conn, back); |
| dri3_copy_area(draw->conn, |
| back->pixmap, |
| draw->drawable, |
| dri3_drawable_gc(draw), |
| x, y, x, y, width, height); |
| dri3_fence_trigger(draw->conn, back); |
| /* Refresh the fake front (if present) after we just damaged the real |
| * front. |
| */ |
| if (draw->have_fake_front && |
| !loader_dri3_blit_image(draw, |
| dri3_front_buffer(draw)->image, |
| back->image, |
| x, y, width, height, |
| x, y, __BLIT_FLAG_FLUSH) && |
| !draw->is_different_gpu) { |
| dri3_fence_reset(draw->conn, dri3_front_buffer(draw)); |
| dri3_copy_area(draw->conn, |
| back->pixmap, |
| dri3_front_buffer(draw)->pixmap, |
| dri3_drawable_gc(draw), |
| x, y, x, y, width, height); |
| dri3_fence_trigger(draw->conn, dri3_front_buffer(draw)); |
| dri3_fence_await(draw->conn, NULL, dri3_front_buffer(draw)); |
| } |
| dri3_fence_await(draw->conn, draw, back); |
| } |
| |
| void |
| loader_dri3_copy_drawable(struct loader_dri3_drawable *draw, |
| xcb_drawable_t dest, |
| xcb_drawable_t src) |
| { |
| loader_dri3_flush(draw, __DRI2_FLUSH_DRAWABLE, __DRI2_THROTTLE_COPYSUBBUFFER); |
| |
| struct loader_dri3_buffer *front = dri3_front_buffer(draw); |
| if (front) |
| dri3_fence_reset(draw->conn, front); |
| |
| dri3_copy_area(draw->conn, |
| src, dest, |
| dri3_drawable_gc(draw), |
| 0, 0, 0, 0, draw->width, draw->height); |
| |
| if (front) { |
| dri3_fence_trigger(draw->conn, front); |
| dri3_fence_await(draw->conn, draw, front); |
| } |
| } |
| |
| void |
| loader_dri3_wait_x(struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *front; |
| |
| if (draw == NULL || !draw->have_fake_front) |
| return; |
| |
| front = dri3_front_buffer(draw); |
| |
| loader_dri3_copy_drawable(draw, front->pixmap, draw->drawable); |
| |
| /* In the psc->is_different_gpu case, the linear buffer has been updated, |
| * but not yet the tiled buffer. |
| * Copy back to the tiled buffer we use for rendering. |
| * Note that we don't need flushing. |
| */ |
| if (draw->is_different_gpu) |
| (void) loader_dri3_blit_image(draw, |
| front->image, |
| front->linear_buffer, |
| 0, 0, front->width, front->height, |
| 0, 0, 0); |
| } |
| |
| void |
| loader_dri3_wait_gl(struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *front; |
| |
| if (draw == NULL || !draw->have_fake_front) |
| return; |
| |
| front = dri3_front_buffer(draw); |
| |
| /* In the psc->is_different_gpu case, we update the linear_buffer |
| * before updating the real front. |
| */ |
| if (draw->is_different_gpu) |
| (void) loader_dri3_blit_image(draw, |
| front->linear_buffer, |
| front->image, |
| 0, 0, front->width, front->height, |
| 0, 0, __BLIT_FLAG_FLUSH); |
| loader_dri3_swapbuffer_barrier(draw); |
| loader_dri3_copy_drawable(draw, draw->drawable, front->pixmap); |
| } |
| |
| /** dri3_flush_present_events |
| * |
| * Process any present events that have been received from the X server |
| */ |
| static void |
| dri3_flush_present_events(struct loader_dri3_drawable *draw) |
| { |
| /* Check to see if any configuration changes have occurred |
| * since we were last invoked |
| */ |
| if (draw->has_event_waiter) |
| return; |
| |
| if (draw->special_event) { |
| xcb_generic_event_t *ev; |
| |
| while ((ev = xcb_poll_for_special_event(draw->conn, |
| draw->special_event)) != NULL) { |
| xcb_present_generic_event_t *ge = (void *) ev; |
| if (!dri3_handle_present_event(draw, ge)) |
| break; |
| } |
| } |
| } |
| |
| /** loader_dri3_swap_buffers_msc |
| * |
| * Make the current back buffer visible using the present extension |
| */ |
| int64_t |
| loader_dri3_swap_buffers_msc(struct loader_dri3_drawable *draw, |
| int64_t target_msc, int64_t divisor, |
| int64_t remainder, unsigned flush_flags, |
| const int *rects, int n_rects, |
| bool force_copy) |
| { |
| struct loader_dri3_buffer *back; |
| int64_t ret = 0; |
| bool wait_for_next_buffer = false; |
| |
| /* GLX spec: |
| * void glXSwapBuffers(Display *dpy, GLXDrawable draw); |
| * This operation is a no-op if draw was created with a non-double-buffered |
| * GLXFBConfig, or if draw is a GLXPixmap. |
| * ... |
| * GLX pixmaps may be created with a config that includes back buffers and |
| * stereoscopic buffers. However, glXSwapBuffers is ignored for these pixmaps. |
| * ... |
| * It is possible to create a pbuffer with back buffers and to swap the |
| * front and back buffers by calling glXSwapBuffers. |
| * |
| * EGL spec: |
| * EGLBoolean eglSwapBuffers(EGLDisplay dpy, EGLSurface surface); |
| * If surface is a back-buffered window surface, then the color buffer is |
| * copied to the native window associated with that surface. If surface is |
| * a single-buffered window, pixmap, or pbuffer surface, eglSwapBuffers has |
| * no effect. |
| * |
| * SwapBuffer effect: |
| * | GLX | EGL | |
| * | window | pixmap | pbuffer | window | pixmap | pbuffer| |
| *-------+--------+--------+---------+--------+--------+--------+ |
| * single| nop | nop | nop | nop | nop | nop | |
| * double| swap | nop | swap | swap | NA | NA | |
| */ |
| if (!draw->have_back || draw->type == LOADER_DRI3_DRAWABLE_PIXMAP) |
| return ret; |
| |
| draw->vtable->flush_drawable(draw, flush_flags); |
| |
| back = dri3_find_back_alloc(draw); |
| /* Could only happen when error case, like display is already closed. */ |
| if (!back) |
| return ret; |
| |
| mtx_lock(&draw->mtx); |
| |
| if (draw->adaptive_sync && !draw->adaptive_sync_active) { |
| set_adaptive_sync_property(draw->conn, draw->drawable, true); |
| draw->adaptive_sync_active = true; |
| } |
| |
| if (draw->is_different_gpu) { |
| /* Update the linear buffer before presenting the pixmap */ |
| (void) loader_dri3_blit_image(draw, |
| back->linear_buffer, |
| back->image, |
| 0, 0, back->width, back->height, |
| 0, 0, __BLIT_FLAG_FLUSH); |
| } |
| |
| /* If we need to preload the new back buffer, remember the source. |
| * The force_copy parameter is used by EGL to attempt to preserve |
| * the back buffer across a call to this function. |
| */ |
| if (draw->swap_method != __DRI_ATTRIB_SWAP_UNDEFINED || force_copy) |
| draw->cur_blit_source = LOADER_DRI3_BACK_ID(draw->cur_back); |
| |
| /* Exchange the back and fake front. Even though the server knows about these |
| * buffers, it has no notion of back and fake front. |
| */ |
| if (draw->have_fake_front) { |
| struct loader_dri3_buffer *tmp; |
| |
| tmp = dri3_front_buffer(draw); |
| draw->buffers[LOADER_DRI3_FRONT_ID] = back; |
| draw->buffers[LOADER_DRI3_BACK_ID(draw->cur_back)] = tmp; |
| |
| if (draw->swap_method == __DRI_ATTRIB_SWAP_COPY || force_copy) |
| draw->cur_blit_source = LOADER_DRI3_FRONT_ID; |
| } |
| |
| dri3_flush_present_events(draw); |
| |
| if (draw->type == LOADER_DRI3_DRAWABLE_WINDOW) { |
| dri3_fence_reset(draw->conn, back); |
| |
| /* Compute when we want the frame shown by taking the last known |
| * successful MSC and adding in a swap interval for each outstanding swap |
| * request. target_msc=divisor=remainder=0 means "Use glXSwapBuffers() |
| * semantic" |
| */ |
| ++draw->send_sbc; |
| if (target_msc == 0 && divisor == 0 && remainder == 0) |
| target_msc = draw->msc + abs(draw->swap_interval) * |
| (draw->send_sbc - draw->recv_sbc); |
| else if (divisor == 0 && remainder > 0) { |
| /* From the GLX_OML_sync_control spec: |
| * "If <divisor> = 0, the swap will occur when MSC becomes |
| * greater than or equal to <target_msc>." |
| * |
| * Note that there's no mention of the remainder. The Present |
| * extension throws BadValue for remainder != 0 with divisor == 0, so |
| * just drop the passed in value. |
| */ |
| remainder = 0; |
| } |
| |
| /* From the GLX_EXT_swap_control spec |
| * and the EGL 1.4 spec (page 53): |
| * |
| * "If <interval> is set to a value of 0, buffer swaps are not |
| * synchronized to a video frame." |
| * |
| * From GLX_EXT_swap_control_tear: |
| * |
| * "If <interval> is negative, the minimum number of video frames |
| * between buffer swaps is the absolute value of <interval>. In this |
| * case, if abs(<interval>) video frames have already passed from |
| * the previous swap when the swap is ready to be performed, the |
| * swap will occur without synchronization to a video frame." |
| * |
| * Implementation note: It is possible to enable triple buffering |
| * behaviour by not using XCB_PRESENT_OPTION_ASYNC, but this should not be |
| * the default. |
| */ |
| uint32_t options = XCB_PRESENT_OPTION_NONE; |
| if (draw->swap_interval <= 0) |
| options |= XCB_PRESENT_OPTION_ASYNC; |
| |
| /* If we need to populate the new back, but need to reuse the back |
| * buffer slot due to lack of local blit capabilities, make sure |
| * the server doesn't flip and we deadlock. |
| */ |
| if (!loader_dri3_have_image_blit(draw) && draw->cur_blit_source != -1) |
| options |= XCB_PRESENT_OPTION_COPY; |
| #ifdef HAVE_DRI3_MODIFIERS |
| if (draw->multiplanes_available) |
| options |= XCB_PRESENT_OPTION_SUBOPTIMAL; |
| #endif |
| back->busy = 1; |
| back->last_swap = draw->send_sbc; |
| |
| if (!draw->region) { |
| draw->region = xcb_generate_id(draw->conn); |
| xcb_xfixes_create_region(draw->conn, draw->region, 0, NULL); |
| } |
| |
| xcb_xfixes_region_t region = 0; |
| xcb_rectangle_t xcb_rects[64]; |
| |
| if (n_rects > 0 && n_rects <= ARRAY_SIZE(xcb_rects)) { |
| for (int i = 0; i < n_rects; i++) { |
| const int *rect = &rects[i * 4]; |
| xcb_rects[i].x = rect[0]; |
| xcb_rects[i].y = draw->height - rect[1] - rect[3]; |
| xcb_rects[i].width = rect[2]; |
| xcb_rects[i].height = rect[3]; |
| } |
| |
| region = draw->region; |
| xcb_xfixes_set_region(draw->conn, region, n_rects, xcb_rects); |
| } |
| |
| xcb_present_pixmap(draw->conn, |
| draw->drawable, |
| back->pixmap, |
| (uint32_t) draw->send_sbc, |
| 0, /* valid */ |
| region, /* update */ |
| 0, /* x_off */ |
| 0, /* y_off */ |
| None, /* target_crtc */ |
| None, |
| back->sync_fence, |
| options, |
| target_msc, |
| divisor, |
| remainder, 0, NULL); |
| } else { |
| /* This can only be reached by double buffered GLXPbuffer. */ |
| assert(draw->type == LOADER_DRI3_DRAWABLE_PBUFFER); |
| /* GLX does not have damage regions. */ |
| assert(n_rects == 0); |
| |
| /* For wait and buffer age usage. */ |
| draw->send_sbc++; |
| draw->recv_sbc = back->last_swap = draw->send_sbc; |
| |
| /* Pixmap is imported as front buffer image when same GPU case, so just |
| * locally blit back buffer image to it is enough. Otherwise front buffer |
| * is a fake one which needs to be synced with pixmap by xserver remotely. |
| */ |
| if (draw->is_different_gpu || |
| !loader_dri3_blit_image(draw, |
| dri3_front_buffer(draw)->image, |
| back->image, |
| 0, 0, draw->width, draw->height, |
| 0, 0, __BLIT_FLAG_FLUSH)) { |
| dri3_copy_area(draw->conn, back->pixmap, |
| draw->drawable, |
| dri3_drawable_gc(draw), |
| 0, 0, 0, 0, draw->width, draw->height); |
| } |
| } |
| |
| ret = (int64_t) draw->send_sbc; |
| |
| /* Schedule a server-side back-preserving blit if necessary. |
| * This happens iff all conditions below are satisfied: |
| * a) We have a fake front, |
| * b) We need to preserve the back buffer, |
| * c) We don't have local blit capabilities. |
| */ |
| if (!loader_dri3_have_image_blit(draw) && draw->cur_blit_source != -1 && |
| draw->cur_blit_source != LOADER_DRI3_BACK_ID(draw->cur_back)) { |
| struct loader_dri3_buffer *new_back = dri3_back_buffer(draw); |
| struct loader_dri3_buffer *src = draw->buffers[draw->cur_blit_source]; |
| |
| dri3_fence_reset(draw->conn, new_back); |
| dri3_copy_area(draw->conn, src->pixmap, |
| new_back->pixmap, |
| dri3_drawable_gc(draw), |
| 0, 0, 0, 0, draw->width, draw->height); |
| dri3_fence_trigger(draw->conn, new_back); |
| new_back->last_swap = src->last_swap; |
| } |
| |
| xcb_flush(draw->conn); |
| if (draw->stamp) |
| ++(*draw->stamp); |
| |
| /* Waiting on a buffer is only sensible if all buffers are in use and the |
| * client doesn't use the buffer age extension. In this case a client is |
| * relying on it receiving back control immediately. |
| * |
| * As waiting on a buffer can at worst make us miss a frame the option has |
| * to be enabled explicitly with the block_on_depleted_buffers DRI option. |
| */ |
| wait_for_next_buffer = draw->cur_num_back == draw->max_num_back && |
| !draw->queries_buffer_age && draw->block_on_depleted_buffers; |
| |
| mtx_unlock(&draw->mtx); |
| |
| draw->ext->flush->invalidate(draw->dri_drawable); |
| |
| /* Clients that use up all available buffers usually regulate their drawing |
| * through swapchain contention backpressure. In such a scenario the client |
| * draws whenever control returns to it. Its event loop is slowed down only |
| * by us waiting on buffers becoming available again. |
| * |
| * By waiting here on a new buffer and only then returning back to the client |
| * we ensure the client begins drawing only when the next buffer is available |
| * and not draw first and then wait a refresh cycle on the next available |
| * buffer to show it. This way we can reduce the latency between what is |
| * being drawn by the client and what is shown on the screen by one frame. |
| */ |
| if (wait_for_next_buffer) |
| dri3_find_back(draw, draw->prefer_back_buffer_reuse); |
| |
| return ret; |
| } |
| |
| int |
| loader_dri3_query_buffer_age(struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *back = dri3_find_back_alloc(draw); |
| int ret = 0; |
| |
| mtx_lock(&draw->mtx); |
| draw->queries_buffer_age = true; |
| if (back && back->last_swap != 0) |
| ret = draw->send_sbc - back->last_swap + 1; |
| mtx_unlock(&draw->mtx); |
| |
| return ret; |
| } |
| |
| /** loader_dri3_open |
| * |
| * Wrapper around xcb_dri3_open |
| */ |
| int |
| loader_dri3_open(xcb_connection_t *conn, |
| xcb_window_t root, |
| uint32_t provider) |
| { |
| xcb_dri3_open_cookie_t cookie; |
| xcb_dri3_open_reply_t *reply; |
| xcb_xfixes_query_version_cookie_t fixes_cookie; |
| xcb_xfixes_query_version_reply_t *fixes_reply; |
| int fd; |
| |
| cookie = xcb_dri3_open(conn, |
| root, |
| provider); |
| |
| reply = xcb_dri3_open_reply(conn, cookie, NULL); |
| |
| if (!reply || reply->nfd != 1) { |
| free(reply); |
| return -1; |
| } |
| |
| fd = xcb_dri3_open_reply_fds(conn, reply)[0]; |
| free(reply); |
| fcntl(fd, F_SETFD, fcntl(fd, F_GETFD) | FD_CLOEXEC); |
| |
| /* let the server know our xfixes level */ |
| fixes_cookie = xcb_xfixes_query_version(conn, |
| XCB_XFIXES_MAJOR_VERSION, |
| XCB_XFIXES_MINOR_VERSION); |
| fixes_reply = xcb_xfixes_query_version_reply(conn, fixes_cookie, NULL); |
| free(fixes_reply); |
| |
| return fd; |
| } |
| |
| static uint32_t |
| dri3_cpp_for_format(uint32_t format) { |
| switch (format) { |
| case __DRI_IMAGE_FORMAT_R8: |
| return 1; |
| case __DRI_IMAGE_FORMAT_RGB565: |
| case __DRI_IMAGE_FORMAT_GR88: |
| return 2; |
| case __DRI_IMAGE_FORMAT_XRGB8888: |
| case __DRI_IMAGE_FORMAT_ARGB8888: |
| case __DRI_IMAGE_FORMAT_ABGR8888: |
| case __DRI_IMAGE_FORMAT_XBGR8888: |
| case __DRI_IMAGE_FORMAT_XRGB2101010: |
| case __DRI_IMAGE_FORMAT_ARGB2101010: |
| case __DRI_IMAGE_FORMAT_XBGR2101010: |
| case __DRI_IMAGE_FORMAT_ABGR2101010: |
| case __DRI_IMAGE_FORMAT_SARGB8: |
| case __DRI_IMAGE_FORMAT_SABGR8: |
| case __DRI_IMAGE_FORMAT_SXRGB8: |
| return 4; |
| case __DRI_IMAGE_FORMAT_ABGR16161616: |
| case __DRI_IMAGE_FORMAT_XBGR16161616: |
| case __DRI_IMAGE_FORMAT_XBGR16161616F: |
| case __DRI_IMAGE_FORMAT_ABGR16161616F: |
| return 8; |
| case __DRI_IMAGE_FORMAT_NONE: |
| default: |
| return 0; |
| } |
| } |
| |
| /* Map format of render buffer to corresponding format for the linear_buffer |
| * used for sharing with the display gpu of a Prime setup (== is_different_gpu). |
| * Usually linear_format == format, except for depth >= 30 formats, where |
| * different gpu vendors have different preferences wrt. color channel ordering. |
| */ |
| static uint32_t |
| dri3_linear_format_for_format(struct loader_dri3_drawable *draw, uint32_t format) |
| { |
| switch (format) { |
| case __DRI_IMAGE_FORMAT_XRGB2101010: |
| case __DRI_IMAGE_FORMAT_XBGR2101010: |
| /* Different preferred formats for different hw */ |
| if (dri3_get_red_mask_for_depth(draw, 30) == 0x3ff) |
| return __DRI_IMAGE_FORMAT_XBGR2101010; |
| else |
| return __DRI_IMAGE_FORMAT_XRGB2101010; |
| |
| case __DRI_IMAGE_FORMAT_ARGB2101010: |
| case __DRI_IMAGE_FORMAT_ABGR2101010: |
| /* Different preferred formats for different hw */ |
| if (dri3_get_red_mask_for_depth(draw, 30) == 0x3ff) |
| return __DRI_IMAGE_FORMAT_ABGR2101010; |
| else |
| return __DRI_IMAGE_FORMAT_ARGB2101010; |
| |
| default: |
| return format; |
| } |
| } |
| |
| /* the DRIimage createImage function takes __DRI_IMAGE_FORMAT codes, while |
| * the createImageFromFds call takes DRM_FORMAT codes. To avoid |
| * complete confusion, just deal in __DRI_IMAGE_FORMAT codes for now and |
| * translate to DRM_FORMAT codes in the call to createImageFromFds |
| */ |
| static int |
| image_format_to_fourcc(int format) |
| { |
| |
| /* Convert from __DRI_IMAGE_FORMAT to DRM_FORMAT (sigh) */ |
| switch (format) { |
| case __DRI_IMAGE_FORMAT_SARGB8: return __DRI_IMAGE_FOURCC_SARGB8888; |
| case __DRI_IMAGE_FORMAT_SABGR8: return __DRI_IMAGE_FOURCC_SABGR8888; |
| case __DRI_IMAGE_FORMAT_SXRGB8: return __DRI_IMAGE_FOURCC_SXRGB8888; |
| case __DRI_IMAGE_FORMAT_RGB565: return DRM_FORMAT_RGB565; |
| case __DRI_IMAGE_FORMAT_XRGB8888: return DRM_FORMAT_XRGB8888; |
| case __DRI_IMAGE_FORMAT_ARGB8888: return DRM_FORMAT_ARGB8888; |
| case __DRI_IMAGE_FORMAT_ABGR8888: return DRM_FORMAT_ABGR8888; |
| case __DRI_IMAGE_FORMAT_XBGR8888: return DRM_FORMAT_XBGR8888; |
| case __DRI_IMAGE_FORMAT_XRGB2101010: return DRM_FORMAT_XRGB2101010; |
| case __DRI_IMAGE_FORMAT_ARGB2101010: return DRM_FORMAT_ARGB2101010; |
| case __DRI_IMAGE_FORMAT_XBGR2101010: return DRM_FORMAT_XBGR2101010; |
| case __DRI_IMAGE_FORMAT_ABGR2101010: return DRM_FORMAT_ABGR2101010; |
| case __DRI_IMAGE_FORMAT_ABGR16161616: return DRM_FORMAT_ABGR16161616; |
| case __DRI_IMAGE_FORMAT_XBGR16161616: return DRM_FORMAT_XBGR16161616; |
| case __DRI_IMAGE_FORMAT_XBGR16161616F: return DRM_FORMAT_XBGR16161616F; |
| case __DRI_IMAGE_FORMAT_ABGR16161616F: return DRM_FORMAT_ABGR16161616F; |
| } |
| return 0; |
| } |
| |
| #ifdef HAVE_DRI3_MODIFIERS |
| static bool |
| has_supported_modifier(struct loader_dri3_drawable *draw, unsigned int format, |
| uint64_t *modifiers, uint32_t count) |
| { |
| uint64_t *supported_modifiers; |
| int32_t supported_modifiers_count; |
| bool found = false; |
| int i, j; |
| |
| if (!draw->ext->image->queryDmaBufModifiers(draw->dri_screen, |
| format, 0, NULL, NULL, |
| &supported_modifiers_count) || |
| supported_modifiers_count == 0) |
| return false; |
| |
| supported_modifiers = malloc(supported_modifiers_count * sizeof(uint64_t)); |
| if (!supported_modifiers) |
| return false; |
| |
| draw->ext->image->queryDmaBufModifiers(draw->dri_screen, format, |
| supported_modifiers_count, |
| supported_modifiers, NULL, |
| &supported_modifiers_count); |
| |
| for (i = 0; !found && i < supported_modifiers_count; i++) { |
| for (j = 0; !found && j < count; j++) { |
| if (supported_modifiers[i] == modifiers[j]) |
| found = true; |
| } |
| } |
| |
| free(supported_modifiers); |
| return found; |
| } |
| #endif |
| |
| /** loader_dri3_alloc_render_buffer |
| * |
| * Use the driver createImage function to construct a __DRIimage, then |
| * get a file descriptor for that and create an X pixmap from that |
| * |
| * Allocate an xshmfence for synchronization |
| */ |
| static struct loader_dri3_buffer * |
| dri3_alloc_render_buffer(struct loader_dri3_drawable *draw, unsigned int format, |
| int width, int height, int depth) |
| { |
| struct loader_dri3_buffer *buffer; |
| __DRIimage *pixmap_buffer = NULL, *linear_buffer_display_gpu = NULL; |
| xcb_pixmap_t pixmap; |
| xcb_sync_fence_t sync_fence; |
| struct xshmfence *shm_fence; |
| int buffer_fds[4], fence_fd; |
| int num_planes = 0; |
| uint64_t *modifiers = NULL; |
| uint32_t count = 0; |
| int i, mod; |
| int ret; |
| |
| /* Create an xshmfence object and |
| * prepare to send that to the X server |
| */ |
| |
| fence_fd = xshmfence_alloc_shm(); |
| if (fence_fd < 0) |
| return NULL; |
| |
| shm_fence = xshmfence_map_shm(fence_fd); |
| if (shm_fence == NULL) |
| goto no_shm_fence; |
| |
| /* Allocate the image from the driver |
| */ |
| buffer = calloc(1, sizeof *buffer); |
| if (!buffer) |
| goto no_buffer; |
| |
| buffer->cpp = dri3_cpp_for_format(format); |
| if (!buffer->cpp) |
| goto no_image; |
| |
| if (!draw->is_different_gpu) { |
| #ifdef HAVE_DRI3_MODIFIERS |
| if (draw->multiplanes_available && |
| draw->ext->image->base.version >= 15 && |
| draw->ext->image->queryDmaBufModifiers && |
| draw->ext->image->createImageWithModifiers) { |
| xcb_dri3_get_supported_modifiers_cookie_t mod_cookie; |
| xcb_dri3_get_supported_modifiers_reply_t *mod_reply; |
| xcb_generic_error_t *error = NULL; |
| |
| mod_cookie = xcb_dri3_get_supported_modifiers(draw->conn, |
| draw->window, |
| depth, buffer->cpp * 8); |
| mod_reply = xcb_dri3_get_supported_modifiers_reply(draw->conn, |
| mod_cookie, |
| &error); |
| if (!mod_reply) |
| goto no_image; |
| |
| if (mod_reply->num_window_modifiers) { |
| count = mod_reply->num_window_modifiers; |
| modifiers = malloc(count * sizeof(uint64_t)); |
| if (!modifiers) { |
| free(mod_reply); |
| goto no_image; |
| } |
| |
| memcpy(modifiers, |
| xcb_dri3_get_supported_modifiers_window_modifiers(mod_reply), |
| count * sizeof(uint64_t)); |
| |
| if (!has_supported_modifier(draw, image_format_to_fourcc(format), |
| modifiers, count)) { |
| free(modifiers); |
| count = 0; |
| modifiers = NULL; |
| } |
| } |
| |
| if (mod_reply->num_screen_modifiers && modifiers == NULL) { |
| count = mod_reply->num_screen_modifiers; |
| modifiers = malloc(count * sizeof(uint64_t)); |
| if (!modifiers) { |
| free(mod_reply); |
| goto no_image; |
| } |
| |
| memcpy(modifiers, |
| xcb_dri3_get_supported_modifiers_screen_modifiers(mod_reply), |
| count * sizeof(uint64_t)); |
| } |
| |
| free(mod_reply); |
| } |
| #endif |
| buffer->image = loader_dri_create_image(draw->dri_screen, draw->ext->image, |
| width, height, format, |
| __DRI_IMAGE_USE_SHARE | |
| __DRI_IMAGE_USE_SCANOUT | |
| __DRI_IMAGE_USE_BACKBUFFER | |
| (draw->is_protected_content ? |
| __DRI_IMAGE_USE_PROTECTED : 0), |
| modifiers, count, buffer); |
| free(modifiers); |
| |
| pixmap_buffer = buffer->image; |
| |
| if (!buffer->image) |
| goto no_image; |
| } else { |
| buffer->image = draw->ext->image->createImage(draw->dri_screen, |
| width, height, |
| format, |
| 0, |
| buffer); |
| |
| if (!buffer->image) |
| goto no_image; |
| |
| /* if driver name is same only then dri_screen_display_gpu is set. |
| * This check is needed because for simplicity render gpu image extension |
| * is also used for display gpu. |
| */ |
| if (draw->dri_screen_display_gpu) { |
| linear_buffer_display_gpu = |
| draw->ext->image->createImage(draw->dri_screen_display_gpu, |
| width, height, |
| dri3_linear_format_for_format(draw, format), |
| __DRI_IMAGE_USE_SHARE | |
| __DRI_IMAGE_USE_LINEAR | |
| __DRI_IMAGE_USE_BACKBUFFER | |
| __DRI_IMAGE_USE_SCANOUT, |
| buffer); |
| pixmap_buffer = linear_buffer_display_gpu; |
| } |
| |
| if (!pixmap_buffer) { |
| buffer->linear_buffer = |
| draw->ext->image->createImage(draw->dri_screen, |
| width, height, |
| dri3_linear_format_for_format(draw, format), |
| __DRI_IMAGE_USE_SHARE | |
| __DRI_IMAGE_USE_LINEAR | |
| __DRI_IMAGE_USE_BACKBUFFER | |
| __DRI_IMAGE_USE_SCANOUT | |
| __DRI_IMAGE_USE_PRIME_BUFFER, |
| buffer); |
| |
| pixmap_buffer = buffer->linear_buffer; |
| if (!buffer->linear_buffer) { |
| goto no_linear_buffer; |
| } |
| } |
| } |
| |
| /* X want some information about the planes, so ask the image for it |
| */ |
| if (!draw->ext->image->queryImage(pixmap_buffer, __DRI_IMAGE_ATTRIB_NUM_PLANES, |
| &num_planes)) |
| num_planes = 1; |
| |
| for (i = 0; i < num_planes; i++) { |
| __DRIimage *image = draw->ext->image->fromPlanar(pixmap_buffer, i, NULL); |
| |
| if (!image) { |
| assert(i == 0); |
| image = pixmap_buffer; |
| } |
| |
| buffer_fds[i] = -1; |
| |
| ret = draw->ext->image->queryImage(image, __DRI_IMAGE_ATTRIB_FD, |
| &buffer_fds[i]); |
| ret &= draw->ext->image->queryImage(image, __DRI_IMAGE_ATTRIB_STRIDE, |
| &buffer->strides[i]); |
| ret &= draw->ext->image->queryImage(image, __DRI_IMAGE_ATTRIB_OFFSET, |
| &buffer->offsets[i]); |
| if (image != pixmap_buffer) |
| draw->ext->image->destroyImage(image); |
| |
| if (!ret) |
| goto no_buffer_attrib; |
| } |
| |
| ret = draw->ext->image->queryImage(pixmap_buffer, |
| __DRI_IMAGE_ATTRIB_MODIFIER_UPPER, &mod); |
| buffer->modifier = (uint64_t) mod << 32; |
| ret &= draw->ext->image->queryImage(pixmap_buffer, |
| __DRI_IMAGE_ATTRIB_MODIFIER_LOWER, &mod); |
| buffer->modifier |= (uint64_t)(mod & 0xffffffff); |
| |
| if (!ret) |
| buffer->modifier = DRM_FORMAT_MOD_INVALID; |
| |
| if (draw->is_different_gpu && draw->dri_screen_display_gpu && |
| linear_buffer_display_gpu) { |
| /* The linear buffer was created in the display GPU's vram, so we |
| * need to make it visible to render GPU |
| */ |
| if (draw->ext->image->base.version >= 20) |
| buffer->linear_buffer = |
| draw->ext->image->createImageFromFds2(draw->dri_screen, |
| width, |
| height, |
| image_format_to_fourcc(format), |
| &buffer_fds[0], num_planes, |
| __DRI_IMAGE_PRIME_LINEAR_BUFFER, |
| &buffer->strides[0], |
| &buffer->offsets[0], |
| buffer); |
| else |
| buffer->linear_buffer = |
| draw->ext->image->createImageFromFds(draw->dri_screen, |
| width, |
| height, |
| image_format_to_fourcc(format), |
| &buffer_fds[0], num_planes, |
| &buffer->strides[0], |
| &buffer->offsets[0], |
| buffer); |
| if (!buffer->linear_buffer) |
| goto no_buffer_attrib; |
| |
| draw->ext->image->destroyImage(linear_buffer_display_gpu); |
| } |
| |
| pixmap = xcb_generate_id(draw->conn); |
| #ifdef HAVE_DRI3_MODIFIERS |
| if (draw->multiplanes_available && |
| buffer->modifier != DRM_FORMAT_MOD_INVALID) { |
| xcb_dri3_pixmap_from_buffers(draw->conn, |
| pixmap, |
| draw->window, |
| num_planes, |
| width, height, |
| buffer->strides[0], buffer->offsets[0], |
| buffer->strides[1], buffer->offsets[1], |
| buffer->strides[2], buffer->offsets[2], |
| buffer->strides[3], buffer->offsets[3], |
| depth, buffer->cpp * 8, |
| buffer->modifier, |
| buffer_fds); |
| } else |
| #endif |
| { |
| xcb_dri3_pixmap_from_buffer(draw->conn, |
| pixmap, |
| draw->drawable, |
| buffer->size, |
| width, height, buffer->strides[0], |
| depth, buffer->cpp * 8, |
| buffer_fds[0]); |
| } |
| |
| xcb_dri3_fence_from_fd(draw->conn, |
| pixmap, |
| (sync_fence = xcb_generate_id(draw->conn)), |
| false, |
| fence_fd); |
| |
| buffer->pixmap = pixmap; |
| buffer->own_pixmap = true; |
| buffer->sync_fence = sync_fence; |
| buffer->shm_fence = shm_fence; |
| buffer->width = width; |
| buffer->height = height; |
| |
| /* Mark the buffer as idle |
| */ |
| dri3_fence_set(buffer); |
| |
| return buffer; |
| |
| no_buffer_attrib: |
| do { |
| if (buffer_fds[i] != -1) |
| close(buffer_fds[i]); |
| } while (--i >= 0); |
| draw->ext->image->destroyImage(pixmap_buffer); |
| no_linear_buffer: |
| if (draw->is_different_gpu) |
| draw->ext->image->destroyImage(buffer->image); |
| no_image: |
| free(buffer); |
| no_buffer: |
| xshmfence_unmap_shm(shm_fence); |
| no_shm_fence: |
| close(fence_fd); |
| return NULL; |
| } |
| |
| static bool |
| dri3_detect_drawable_is_window(struct loader_dri3_drawable *draw) |
| { |
| /* Try to select for input on the window. |
| * |
| * If the drawable is a window, this will get our events |
| * delivered. |
| * |
| * Otherwise, we'll get a BadWindow error back from this request which |
| * will let us know that the drawable is a pixmap instead. |
| */ |
| |
| xcb_void_cookie_t cookie = |
| xcb_present_select_input_checked(draw->conn, draw->eid, draw->drawable, |
| XCB_PRESENT_EVENT_MASK_CONFIGURE_NOTIFY | |
| XCB_PRESENT_EVENT_MASK_COMPLETE_NOTIFY | |
| XCB_PRESENT_EVENT_MASK_IDLE_NOTIFY); |
| |
| /* Check to see if our select input call failed. If it failed with a |
| * BadWindow error, then assume the drawable is a pixmap. |
| */ |
| xcb_generic_error_t *error = xcb_request_check(draw->conn, cookie); |
| |
| if (error) { |
| if (error->error_code != BadWindow) { |
| free(error); |
| return false; |
| } |
| free(error); |
| |
| /* pixmap can't get here, see driFetchDrawable(). */ |
| draw->type = LOADER_DRI3_DRAWABLE_PBUFFER; |
| return true; |
| } |
| |
| draw->type = LOADER_DRI3_DRAWABLE_WINDOW; |
| return true; |
| } |
| |
| static bool |
| dri3_setup_present_event(struct loader_dri3_drawable *draw) |
| { |
| /* No need to setup for pixmap drawable. */ |
| if (draw->type == LOADER_DRI3_DRAWABLE_PIXMAP || |
| draw->type == LOADER_DRI3_DRAWABLE_PBUFFER) |
| return true; |
| |
| draw->eid = xcb_generate_id(draw->conn); |
| |
| if (draw->type == LOADER_DRI3_DRAWABLE_WINDOW) { |
| xcb_present_select_input(draw->conn, draw->eid, draw->drawable, |
| XCB_PRESENT_EVENT_MASK_CONFIGURE_NOTIFY | |
| XCB_PRESENT_EVENT_MASK_COMPLETE_NOTIFY | |
| XCB_PRESENT_EVENT_MASK_IDLE_NOTIFY); |
| } else { |
| assert(draw->type == LOADER_DRI3_DRAWABLE_UNKNOWN); |
| |
| if (!dri3_detect_drawable_is_window(draw)) |
| return false; |
| |
| if (draw->type != LOADER_DRI3_DRAWABLE_WINDOW) |
| return true; |
| } |
| |
| /* Create an XCB event queue to hold present events outside of the usual |
| * application event queue |
| */ |
| draw->special_event = xcb_register_for_special_xge(draw->conn, |
| &xcb_present_id, |
| draw->eid, |
| draw->stamp); |
| return true; |
| } |
| |
| /** loader_dri3_update_drawable |
| * |
| * Called the first time we use the drawable and then |
| * after we receive present configure notify events to |
| * track the geometry of the drawable |
| */ |
| static int |
| dri3_update_drawable(struct loader_dri3_drawable *draw) |
| { |
| mtx_lock(&draw->mtx); |
| if (draw->first_init) { |
| xcb_get_geometry_cookie_t geom_cookie; |
| xcb_get_geometry_reply_t *geom_reply; |
| xcb_window_t root_win; |
| |
| draw->first_init = false; |
| |
| if (!dri3_setup_present_event(draw)) { |
| mtx_unlock(&draw->mtx); |
| return false; |
| } |
| |
| geom_cookie = xcb_get_geometry(draw->conn, draw->drawable); |
| |
| geom_reply = xcb_get_geometry_reply(draw->conn, geom_cookie, NULL); |
| |
| if (!geom_reply) { |
| mtx_unlock(&draw->mtx); |
| return false; |
| } |
| draw->width = geom_reply->width; |
| draw->height = geom_reply->height; |
| draw->depth = geom_reply->depth; |
| draw->vtable->set_drawable_size(draw, draw->width, draw->height); |
| root_win = geom_reply->root; |
| |
| free(geom_reply); |
| |
| if (draw->type != LOADER_DRI3_DRAWABLE_WINDOW) |
| draw->window = root_win; |
| else |
| draw->window = draw->drawable; |
| } |
| dri3_flush_present_events(draw); |
| mtx_unlock(&draw->mtx); |
| return true; |
| } |
| |
| __DRIimage * |
| loader_dri3_create_image(xcb_connection_t *c, |
| xcb_dri3_buffer_from_pixmap_reply_t *bp_reply, |
| unsigned int format, |
| __DRIscreen *dri_screen, |
| const __DRIimageExtension *image, |
| void *loaderPrivate) |
| { |
| int *fds; |
| __DRIimage *image_planar, *ret; |
| int stride, offset; |
| |
| /* Get an FD for the pixmap object |
| */ |
| fds = xcb_dri3_buffer_from_pixmap_reply_fds(c, bp_reply); |
| |
| stride = bp_reply->stride; |
| offset = 0; |
| |
| /* createImageFromFds creates a wrapper __DRIimage structure which |
| * can deal with multiple planes for things like Yuv images. So, once |
| * we've gotten the planar wrapper, pull the single plane out of it and |
| * discard the wrapper. |
| */ |
| image_planar = image->createImageFromFds(dri_screen, |
| bp_reply->width, |
| bp_reply->height, |
| image_format_to_fourcc(format), |
| fds, 1, |
| &stride, &offset, loaderPrivate); |
| close(fds[0]); |
| if (!image_planar) |
| return NULL; |
| |
| ret = image->fromPlanar(image_planar, 0, loaderPrivate); |
| |
| if (!ret) |
| ret = image_planar; |
| else |
| image->destroyImage(image_planar); |
| |
| return ret; |
| } |
| |
| #ifdef HAVE_DRI3_MODIFIERS |
| __DRIimage * |
| loader_dri3_create_image_from_buffers(xcb_connection_t *c, |
| xcb_dri3_buffers_from_pixmap_reply_t *bp_reply, |
| unsigned int format, |
| __DRIscreen *dri_screen, |
| const __DRIimageExtension *image, |
| void *loaderPrivate) |
| { |
| __DRIimage *ret; |
| int *fds; |
| uint32_t *strides_in, *offsets_in; |
| int strides[4], offsets[4]; |
| unsigned error; |
| int i; |
| |
| if (bp_reply->nfd > 4) |
| return NULL; |
| |
| fds = xcb_dri3_buffers_from_pixmap_reply_fds(c, bp_reply); |
| strides_in = xcb_dri3_buffers_from_pixmap_strides(bp_reply); |
| offsets_in = xcb_dri3_buffers_from_pixmap_offsets(bp_reply); |
| for (i = 0; i < bp_reply->nfd; i++) { |
| strides[i] = strides_in[i]; |
| offsets[i] = offsets_in[i]; |
| } |
| |
| ret = image->createImageFromDmaBufs2(dri_screen, |
| bp_reply->width, |
| bp_reply->height, |
| image_format_to_fourcc(format), |
| bp_reply->modifier, |
| fds, bp_reply->nfd, |
| strides, offsets, |
| 0, 0, 0, 0, /* UNDEFINED */ |
| &error, loaderPrivate); |
| |
| for (i = 0; i < bp_reply->nfd; i++) |
| close(fds[i]); |
| |
| return ret; |
| } |
| #endif |
| |
| /** dri3_get_pixmap_buffer |
| * |
| * Get the DRM object for a pixmap from the X server and |
| * wrap that with a __DRIimage structure using createImageFromFds |
| */ |
| static struct loader_dri3_buffer * |
| dri3_get_pixmap_buffer(__DRIdrawable *driDrawable, unsigned int format, |
| enum loader_dri3_buffer_type buffer_type, |
| struct loader_dri3_drawable *draw) |
| { |
| int buf_id = loader_dri3_pixmap_buf_id(buffer_type); |
| struct loader_dri3_buffer *buffer = draw->buffers[buf_id]; |
| xcb_drawable_t pixmap; |
| xcb_sync_fence_t sync_fence; |
| struct xshmfence *shm_fence; |
| int width; |
| int height; |
| int fence_fd; |
| __DRIscreen *cur_screen; |
| |
| if (buffer) |
| return buffer; |
| |
| pixmap = draw->drawable; |
| |
| buffer = calloc(1, sizeof *buffer); |
| if (!buffer) |
| goto no_buffer; |
| |
| fence_fd = xshmfence_alloc_shm(); |
| if (fence_fd < 0) |
| goto no_fence; |
| shm_fence = xshmfence_map_shm(fence_fd); |
| if (shm_fence == NULL) { |
| close (fence_fd); |
| goto no_fence; |
| } |
| |
| /* Get the currently-bound screen or revert to using the drawable's screen if |
| * no contexts are currently bound. The latter case is at least necessary for |
| * obs-studio, when using Window Capture (Xcomposite) as a Source. |
| */ |
| cur_screen = draw->vtable->get_dri_screen(); |
| if (!cur_screen) { |
| cur_screen = draw->dri_screen; |
| } |
| |
| xcb_dri3_fence_from_fd(draw->conn, |
| pixmap, |
| (sync_fence = xcb_generate_id(draw->conn)), |
| false, |
| fence_fd); |
| #ifdef HAVE_DRI3_MODIFIERS |
| if (draw->multiplanes_available && |
| draw->ext->image->base.version >= 15 && |
| draw->ext->image->createImageFromDmaBufs2) { |
| xcb_dri3_buffers_from_pixmap_cookie_t bps_cookie; |
| xcb_dri3_buffers_from_pixmap_reply_t *bps_reply; |
| |
| bps_cookie = xcb_dri3_buffers_from_pixmap(draw->conn, pixmap); |
| bps_reply = xcb_dri3_buffers_from_pixmap_reply(draw->conn, bps_cookie, |
| NULL); |
| if (!bps_reply) |
| goto no_image; |
| buffer->image = |
| loader_dri3_create_image_from_buffers(draw->conn, bps_reply, format, |
| cur_screen, draw->ext->image, |
| buffer); |
| width = bps_reply->width; |
| height = bps_reply->height; |
| free(bps_reply); |
| } else |
| #endif |
| { |
| xcb_dri3_buffer_from_pixmap_cookie_t bp_cookie; |
| xcb_dri3_buffer_from_pixmap_reply_t *bp_reply; |
| |
| bp_cookie = xcb_dri3_buffer_from_pixmap(draw->conn, pixmap); |
| bp_reply = xcb_dri3_buffer_from_pixmap_reply(draw->conn, bp_cookie, NULL); |
| if (!bp_reply) |
| goto no_image; |
| |
| buffer->image = loader_dri3_create_image(draw->conn, bp_reply, format, |
| cur_screen, draw->ext->image, |
| buffer); |
| width = bp_reply->width; |
| height = bp_reply->height; |
| free(bp_reply); |
| } |
| |
| if (!buffer->image) |
| goto no_image; |
| |
| buffer->pixmap = pixmap; |
| buffer->own_pixmap = false; |
| buffer->width = width; |
| buffer->height = height; |
| buffer->shm_fence = shm_fence; |
| buffer->sync_fence = sync_fence; |
| |
| draw->buffers[buf_id] = buffer; |
| |
| return buffer; |
| |
| no_image: |
| xcb_sync_destroy_fence(draw->conn, sync_fence); |
| xshmfence_unmap_shm(shm_fence); |
| no_fence: |
| free(buffer); |
| no_buffer: |
| return NULL; |
| } |
| |
| /** dri3_get_buffer |
| * |
| * Find a front or back buffer, allocating new ones as necessary |
| */ |
| static struct loader_dri3_buffer * |
| dri3_get_buffer(__DRIdrawable *driDrawable, |
| unsigned int format, |
| enum loader_dri3_buffer_type buffer_type, |
| struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *buffer; |
| bool fence_await = buffer_type == loader_dri3_buffer_back; |
| int buf_id; |
| |
| if (buffer_type == loader_dri3_buffer_back) { |
| draw->back_format = format; |
| |
| buf_id = dri3_find_back(draw, !draw->prefer_back_buffer_reuse); |
| |
| if (buf_id < 0) |
| return NULL; |
| } else { |
| buf_id = LOADER_DRI3_FRONT_ID; |
| } |
| |
| buffer = draw->buffers[buf_id]; |
| |
| /* Allocate a new buffer if there isn't an old one, if that |
| * old one is the wrong size, or if it's suboptimal |
| */ |
| if (!buffer || buffer->width != draw->width || |
| buffer->height != draw->height || |
| buffer->reallocate) { |
| struct loader_dri3_buffer *new_buffer; |
| |
| /* Allocate the new buffers |
| */ |
| new_buffer = dri3_alloc_render_buffer(draw, |
| format, |
| draw->width, |
| draw->height, |
| draw->depth); |
| if (!new_buffer) |
| return NULL; |
| |
| /* When resizing, copy the contents of the old buffer, waiting for that |
| * copy to complete using our fences before proceeding |
| */ |
| if ((buffer_type == loader_dri3_buffer_back || |
| (buffer_type == loader_dri3_buffer_front && draw->have_fake_front)) |
| && buffer) { |
| |
| /* Fill the new buffer with data from an old buffer */ |
| if (!loader_dri3_blit_image(draw, |
| new_buffer->image, |
| buffer->image, |
| 0, 0, |
| MIN2(buffer->width, new_buffer->width), |
| MIN2(buffer->height, new_buffer->height), |
| 0, 0, 0) && |
| !buffer->linear_buffer) { |
| dri3_fence_reset(draw->conn, new_buffer); |
| dri3_copy_area(draw->conn, |
| buffer->pixmap, |
| new_buffer->pixmap, |
| dri3_drawable_gc(draw), |
| 0, 0, 0, 0, |
| draw->width, draw->height); |
| dri3_fence_trigger(draw->conn, new_buffer); |
| fence_await = true; |
| } |
| dri3_free_render_buffer(draw, buffer); |
| } else if (buffer_type == loader_dri3_buffer_front) { |
| /* Fill the new fake front with data from a real front */ |
| loader_dri3_swapbuffer_barrier(draw); |
| dri3_fence_reset(draw->conn, new_buffer); |
| dri3_copy_area(draw->conn, |
| draw->drawable, |
| new_buffer->pixmap, |
| dri3_drawable_gc(draw), |
| 0, 0, 0, 0, |
| draw->width, draw->height); |
| dri3_fence_trigger(draw->conn, new_buffer); |
| |
| if (new_buffer->linear_buffer) { |
| dri3_fence_await(draw->conn, draw, new_buffer); |
| (void) loader_dri3_blit_image(draw, |
| new_buffer->image, |
| new_buffer->linear_buffer, |
| 0, 0, draw->width, draw->height, |
| 0, 0, 0); |
| } else |
| fence_await = true; |
| } |
| buffer = new_buffer; |
| draw->buffers[buf_id] = buffer; |
| } |
| |
| if (fence_await) |
| dri3_fence_await(draw->conn, draw, buffer); |
| |
| /* |
| * Do we need to preserve the content of a previous buffer? |
| * |
| * Note that this blit is needed only to avoid a wait for a buffer that |
| * is currently in the flip chain or being scanned out from. That's really |
| * a tradeoff. If we're ok with the wait we can reduce the number of back |
| * buffers to 1 for SWAP_EXCHANGE, and 1 for SWAP_COPY, |
| * but in the latter case we must disallow page-flipping. |
| */ |
| if (buffer_type == loader_dri3_buffer_back && |
| draw->cur_blit_source != -1 && |
| draw->buffers[draw->cur_blit_source] && |
| buffer != draw->buffers[draw->cur_blit_source]) { |
| |
| struct loader_dri3_buffer *source = draw->buffers[draw->cur_blit_source]; |
| |
| /* Avoid flushing here. Will propably do good for tiling hardware. */ |
| (void) loader_dri3_blit_image(draw, |
| buffer->image, |
| source->image, |
| 0, 0, draw->width, draw->height, |
| 0, 0, 0); |
| buffer->last_swap = source->last_swap; |
| draw->cur_blit_source = -1; |
| } |
| /* Return the requested buffer */ |
| return buffer; |
| } |
| |
| /** dri3_free_buffers |
| * |
| * Free the front bufffer or all of the back buffers. Used |
| * when the application changes which buffers it needs |
| */ |
| static void |
| dri3_free_buffers(__DRIdrawable *driDrawable, |
| enum loader_dri3_buffer_type buffer_type, |
| struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *buffer; |
| int first_id; |
| int n_id; |
| int buf_id; |
| |
| switch (buffer_type) { |
| case loader_dri3_buffer_back: |
| first_id = LOADER_DRI3_BACK_ID(0); |
| n_id = LOADER_DRI3_MAX_BACK; |
| draw->cur_blit_source = -1; |
| break; |
| case loader_dri3_buffer_front: |
| first_id = LOADER_DRI3_FRONT_ID; |
| /* Don't free a fake front holding new backbuffer content. */ |
| n_id = (draw->cur_blit_source == LOADER_DRI3_FRONT_ID) ? 0 : 1; |
| break; |
| default: |
| unreachable("unhandled buffer_type"); |
| } |
| |
| for (buf_id = first_id; buf_id < first_id + n_id; buf_id++) { |
| buffer = draw->buffers[buf_id]; |
| if (buffer) { |
| dri3_free_render_buffer(draw, buffer); |
| draw->buffers[buf_id] = NULL; |
| } |
| } |
| } |
| |
| /** loader_dri3_get_buffers |
| * |
| * The published buffer allocation API. |
| * Returns all of the necessary buffers, allocating |
| * as needed. |
| */ |
| int |
| loader_dri3_get_buffers(__DRIdrawable *driDrawable, |
| unsigned int format, |
| uint32_t *stamp, |
| void *loaderPrivate, |
| uint32_t buffer_mask, |
| struct __DRIimageList *buffers) |
| { |
| struct loader_dri3_drawable *draw = loaderPrivate; |
| struct loader_dri3_buffer *front, *back; |
| int buf_id; |
| |
| buffers->image_mask = 0; |
| buffers->front = NULL; |
| buffers->back = NULL; |
| |
| front = NULL; |
| back = NULL; |
| |
| if (!dri3_update_drawable(draw)) |
| return false; |
| |
| dri3_update_max_num_back(draw); |
| |
| /* Free no longer needed back buffers */ |
| for (buf_id = draw->cur_num_back; buf_id < LOADER_DRI3_MAX_BACK; buf_id++) { |
| if (draw->cur_blit_source != buf_id && draw->buffers[buf_id]) { |
| dri3_free_render_buffer(draw, draw->buffers[buf_id]); |
| draw->buffers[buf_id] = NULL; |
| } |
| } |
| |
| /* pixmaps always have front buffers. |
| * Exchange swaps also mandate fake front buffers. |
| */ |
| if (draw->type != LOADER_DRI3_DRAWABLE_WINDOW || |
| draw->swap_method == __DRI_ATTRIB_SWAP_EXCHANGE) |
| buffer_mask |= __DRI_IMAGE_BUFFER_FRONT; |
| |
| if (buffer_mask & __DRI_IMAGE_BUFFER_FRONT) { |
| /* All pixmaps are owned by the server gpu. |
| * When we use a different gpu, we can't use the pixmap |
| * as buffer since it is potentially tiled a way |
| * our device can't understand. In this case, use |
| * a fake front buffer. Hopefully the pixmap |
| * content will get synced with the fake front |
| * buffer. |
| */ |
| if (draw->type != LOADER_DRI3_DRAWABLE_WINDOW && !draw->is_different_gpu) |
| front = dri3_get_pixmap_buffer(driDrawable, |
| format, |
| loader_dri3_buffer_front, |
| draw); |
| else |
| front = dri3_get_buffer(driDrawable, |
| format, |
| loader_dri3_buffer_front, |
| draw); |
| |
| if (!front) |
| return false; |
| } else { |
| dri3_free_buffers(driDrawable, loader_dri3_buffer_front, draw); |
| draw->have_fake_front = 0; |
| } |
| |
| if (buffer_mask & __DRI_IMAGE_BUFFER_BACK) { |
| back = dri3_get_buffer(driDrawable, |
| format, |
| loader_dri3_buffer_back, |
| draw); |
| if (!back) |
| return false; |
| draw->have_back = 1; |
| } else { |
| dri3_free_buffers(driDrawable, loader_dri3_buffer_back, draw); |
| draw->have_back = 0; |
| } |
| |
| if (front) { |
| buffers->image_mask |= __DRI_IMAGE_BUFFER_FRONT; |
| buffers->front = front->image; |
| draw->have_fake_front = |
| draw->is_different_gpu || |
| draw->type == LOADER_DRI3_DRAWABLE_WINDOW; |
| } |
| |
| if (back) { |
| buffers->image_mask |= __DRI_IMAGE_BUFFER_BACK; |
| buffers->back = back->image; |
| } |
| |
| draw->stamp = stamp; |
| |
| return true; |
| } |
| |
| /** loader_dri3_update_drawable_geometry |
| * |
| * Get the current drawable geometry. |
| */ |
| void |
| loader_dri3_update_drawable_geometry(struct loader_dri3_drawable *draw) |
| { |
| xcb_get_geometry_cookie_t geom_cookie; |
| xcb_get_geometry_reply_t *geom_reply; |
| |
| geom_cookie = xcb_get_geometry(draw->conn, draw->drawable); |
| |
| geom_reply = xcb_get_geometry_reply(draw->conn, geom_cookie, NULL); |
| |
| if (geom_reply) { |
| draw->width = geom_reply->width; |
| draw->height = geom_reply->height; |
| draw->vtable->set_drawable_size(draw, draw->width, draw->height); |
| draw->ext->flush->invalidate(draw->dri_drawable); |
| |
| free(geom_reply); |
| } |
| } |
| |
| void |
| loader_dri3_init_screen_resources(struct loader_dri3_screen_resources *res, |
| xcb_connection_t *conn, |
| xcb_screen_t *screen) |
| { |
| res->conn = conn; |
| res->screen = screen; |
| res->crtcs = NULL; |
| |
| mtx_init(&res->mtx, mtx_plain); |
| } |
| |
| void |
| loader_dri3_destroy_screen_resources(struct loader_dri3_screen_resources *res) |
| { |
| mtx_destroy(&res->mtx); |
| } |
| |
| static unsigned |
| gcd_u32(unsigned a, unsigned b) |
| { |
| assert(a > 0 || b > 0); |
| |
| while (b != 0) { |
| unsigned remainder = a % b; |
| a = b; |
| b = remainder; |
| } |
| |
| return a; |
| } |
| |
| static void |
| calculate_refresh_rate(const xcb_randr_mode_info_t *mode, |
| unsigned *numerator, unsigned *denominator) |
| { |
| unsigned vtotal = mode->vtotal; |
| |
| /* Double-scan doubles the number of lines */ |
| if (mode->mode_flags & XCB_RANDR_MODE_FLAG_DOUBLE_SCAN) |
| vtotal *= 2; |
| |
| /* Interlace splits the frame into two fields; typically the monitor |
| * reports field rate. |
| */ |
| if (mode->mode_flags & XCB_RANDR_MODE_FLAG_INTERLACE) |
| vtotal /= 2; |
| |
| uint32_t dots = mode->htotal * vtotal; |
| |
| if (dots == 0) { |
| *numerator = 0; |
| *denominator = 1; |
| } else { |
| uint32_t gcd = gcd_u32(mode->dot_clock, dots); |
| |
| *numerator = mode->dot_clock / gcd; |
| *denominator = dots / gcd; |
| } |
| } |
| |
| bool |
| loader_dri3_update_screen_resources(struct loader_dri3_screen_resources *res) |
| { |
| xcb_randr_get_crtc_info_cookie_t *crtc_cookies; |
| |
| /* If we have cached screen resources information, check each CRTC to |
| * see if it's up to date. Ideally, we'd watch PresentConfigureNotify |
| * events on the root window to see if something changed, but those only |
| * fire if the geometry changes. It misses CRTC changes which only |
| * alter the refresh rate. We also can't watch RandR events internally |
| * because they aren't XGE events. So, we just check every CRTC for now. |
| */ |
| bool config_unchanged = res->crtcs != NULL; |
| |
| crtc_cookies = malloc(res->num_crtcs * sizeof(*crtc_cookies)); |
| |
| for (unsigned c = 0; c < res->num_crtcs; c++) { |
| crtc_cookies[c] = |
| xcb_randr_get_crtc_info_unchecked(res->conn, res->crtcs[c].id, |
| res->config_timestamp); |
| } |
| |
| for (unsigned c = 0; c < res->num_crtcs; c++) { |
| xcb_randr_get_crtc_info_reply_t *reply = |
| xcb_randr_get_crtc_info_reply(res->conn, crtc_cookies[c], NULL); |
| |
| /* Although randrproto 1.4.0 says that RRGetCrtcInfo is supposed to |
| * return InvalidConfigTime if config_timestamp is out of date, the |
| * implementation in xserver as of 21.x doesn't actually do so. To |
| * detect changes in refresh rate, we check the returned timestamp |
| * on each tracked CRTC. |
| */ |
| if (!reply || |
| reply->status == XCB_RANDR_SET_CONFIG_INVALID_CONFIG_TIME || |
| reply->timestamp != res->crtcs[c].timestamp) { |
| config_unchanged = false; |
| /* continue to consume all replies */ |
| } |
| |
| free(reply); |
| } |
| |
| free(crtc_cookies); |
| |
| if (config_unchanged) |
| return false; |
| |
| /* Do RRGetScreenResourcesCurrent to query the list of CRTCs and modes, |
| * then RRGetCrtcInfo on each CRTC to determine what mode each uses, and |
| * use the mode to calculate the refresh rate. |
| */ |
| mtx_lock(&res->mtx); |
| |
| xcb_randr_get_screen_resources_current_cookie_t cookie = |
| xcb_randr_get_screen_resources_current_unchecked(res->conn, |
| res->screen->root); |
| xcb_randr_get_screen_resources_current_reply_t *reply = |
| xcb_randr_get_screen_resources_current_reply(res->conn, cookie, NULL); |
| |
| xcb_randr_crtc_t *new_crtcs = |
| xcb_randr_get_screen_resources_current_crtcs(reply); |
| |
| xcb_randr_mode_info_t *new_modes = |
| xcb_randr_get_screen_resources_current_modes(reply); |
| |
| res->config_timestamp = reply->config_timestamp; |
| |
| free(res->crtcs); |
| res->crtcs = calloc(reply->num_crtcs, sizeof(*res->crtcs)); |
| |
| crtc_cookies = malloc(reply->num_crtcs * sizeof(*crtc_cookies)); |
| |
| for (unsigned c = 0; c < reply->num_crtcs; c++) { |
| crtc_cookies[c] = |
| xcb_randr_get_crtc_info_unchecked(res->conn, new_crtcs[c], |
| res->config_timestamp); |
| } |
| |
| unsigned i = 0; |
| for (unsigned c = 0; c < reply->num_crtcs; c++) { |
| xcb_randr_get_crtc_info_reply_t *crtc_info = |
| xcb_randr_get_crtc_info_reply(res->conn, crtc_cookies[c], NULL); |
| |
| if (!crtc_info || crtc_info->mode == XCB_NONE) |
| continue; |
| |
| res->crtcs[i].id = new_crtcs[c]; |
| res->crtcs[i].timestamp = crtc_info->timestamp; |
| res->crtcs[i].x = crtc_info->x; |
| res->crtcs[i].y = crtc_info->y; |
| res->crtcs[i].width = crtc_info->width; |
| res->crtcs[i].height = crtc_info->height; |
| |
| for (int m = 0; m < reply->num_modes; m++) { |
| if (new_modes[m].id == crtc_info->mode) { |
| calculate_refresh_rate(&new_modes[m], |
| &res->crtcs[i].refresh_numerator, |
| &res->crtcs[i].refresh_denominator); |
| break; |
| } |
| } |
| |
| i++; |
| free(crtc_info); |
| } |
| |
| res->num_crtcs = i; |
| |
| free(crtc_cookies); |
| free(reply); |
| |
| mtx_unlock(&res->mtx); |
| return true; |
| } |
| |
| /** |
| * Make sure the server has flushed all pending swap buffers to hardware |
| * for this drawable. Ideally we'd want to send an X protocol request to |
| * have the server block our connection until the swaps are complete. That |
| * would avoid the potential round-trip here. |
| */ |
| void |
| loader_dri3_swapbuffer_barrier(struct loader_dri3_drawable *draw) |
| { |
| int64_t ust, msc, sbc; |
| |
| (void) loader_dri3_wait_for_sbc(draw, 0, &ust, &msc, &sbc); |
| } |
| |
| /** |
| * Perform any cleanup associated with a close screen operation. |
| * \param dri_screen[in,out] Pointer to __DRIscreen about to be closed. |
| * |
| * This function destroys the screen's cached swap context if any. |
| */ |
| void |
| loader_dri3_close_screen(__DRIscreen *dri_screen) |
| { |
| mtx_lock(&blit_context.mtx); |
| if (blit_context.ctx && blit_context.cur_screen == dri_screen) { |
| blit_context.core->destroyContext(blit_context.ctx); |
| blit_context.ctx = NULL; |
| } |
| mtx_unlock(&blit_context.mtx); |
| } |
| |
| /** |
| * Find a backbuffer slot - potentially allocating a back buffer |
| * |
| * \param draw[in,out] Pointer to the drawable for which to find back. |
| * \return Pointer to a new back buffer or NULL if allocation failed or was |
| * not mandated. |
| * |
| * Find a potentially new back buffer, and if it's not been allocated yet and |
| * in addition needs initializing, then try to allocate and initialize it. |
| */ |
| #include <stdio.h> |
| static struct loader_dri3_buffer * |
| dri3_find_back_alloc(struct loader_dri3_drawable *draw) |
| { |
| struct loader_dri3_buffer *back; |
| int id; |
| |
| id = dri3_find_back(draw, false); |
| if (id < 0) |
| return NULL; |
| |
| back = draw->buffers[id]; |
| /* Allocate a new back if we haven't got one */ |
| if (!back && draw->back_format != __DRI_IMAGE_FORMAT_NONE && |
| dri3_update_drawable(draw)) |
| back = dri3_alloc_render_buffer(draw, draw->back_format, |
| draw->width, draw->height, draw->depth); |
| |
| if (!back) |
| return NULL; |
| |
| draw->buffers[id] = back; |
| |
| /* If necessary, prefill the back with data according to swap_method mode. */ |
| if (draw->cur_blit_source != -1 && |
| draw->buffers[draw->cur_blit_source] && |
| back != draw->buffers[draw->cur_blit_source]) { |
| struct loader_dri3_buffer *source = draw->buffers[draw->cur_blit_source]; |
| |
| dri3_fence_await(draw->conn, draw, source); |
| dri3_fence_await(draw->conn, draw, back); |
| (void) loader_dri3_blit_image(draw, |
| back->image, |
| source->image, |
| 0, 0, draw->width, draw->height, |
| 0, 0, 0); |
| back->last_swap = source->last_swap; |
| draw->cur_blit_source = -1; |
| } |
| |
| return back; |
| } |