X-Git-Url: http://git.demorecorder.com/cgi-bin/gitweb.cgi?a=blobdiff_plain;f=src%2Fxcb_out.c;h=dc42954676d05019b39b54ff539dfafdb0c1ebe4;hb=c3cfa04b09ff042fed731b17dac3e40de1a10187;hp=367013a9c025b768ade0909f57caa75a708cb553;hpb=29f9fe0fc805a1ec6860f167a45664cc1cf0c769;p=free-sw%2Fxcb%2Flibxcb diff --git a/src/xcb_out.c b/src/xcb_out.c index 367013a..dc42954 100644 --- a/src/xcb_out.c +++ b/src/xcb_out.c @@ -25,57 +25,168 @@ /* Stuff that sends stuff to the server. */ +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + #include #include #include #include -#include #include "xcb.h" #include "xcbext.h" #include "xcbint.h" -#include "extensions/bigreq.h" +#include "bigreq.h" -static int force_sequence_wrap(XCBConnection *c) +static inline void send_request(xcb_connection_t *c, int isvoid, enum workarounds workaround, int flags, struct iovec *vector, int count) { - int ret = 1; - if((c->out.request - c->in.request_read) > 65530) + if(c->has_error) + return; + + ++c->out.request; + if(!isvoid) + c->in.request_expected = c->out.request; + if(workaround != WORKAROUND_NONE || flags != 0) + _xcb_in_expect_reply(c, c->out.request, workaround, flags); + + while(count && c->out.queue_len + vector[0].iov_len <= sizeof(c->out.queue)) { - pthread_mutex_unlock(&c->iolock); - ret = XCBSync(c, 0); - pthread_mutex_lock(&c->iolock); + memcpy(c->out.queue + c->out.queue_len, vector[0].iov_base, vector[0].iov_len); + c->out.queue_len += vector[0].iov_len; + vector[0].iov_base = (char *) vector[0].iov_base + vector[0].iov_len; + vector[0].iov_len = 0; + ++vector, --count; + } + if(!count) + return; + + --vector, ++count; + vector[0].iov_base = c->out.queue; + vector[0].iov_len = c->out.queue_len; + c->out.queue_len = 0; + _xcb_out_send(c, vector, count); +} + +static void send_sync(xcb_connection_t *c) +{ + static const union { + struct { + uint8_t major; + uint8_t pad; + uint16_t len; + } fields; + uint32_t packet; + } sync_req = { { /* GetInputFocus */ 43, 0, 1 } }; + struct iovec vector[2]; + vector[1].iov_base = (char *) &sync_req; + vector[1].iov_len = sizeof(sync_req); + send_request(c, 0, WORKAROUND_NONE, XCB_REQUEST_DISCARD_REPLY, vector + 1, 1); +} + +static void get_socket_back(xcb_connection_t *c) +{ + while(c->out.return_socket && c->out.socket_moving) + pthread_cond_wait(&c->out.socket_cond, &c->iolock); + if(!c->out.return_socket) + return; + + c->out.socket_moving = 1; + pthread_mutex_unlock(&c->iolock); + c->out.return_socket(c->out.socket_closure); + pthread_mutex_lock(&c->iolock); + c->out.socket_moving = 0; + + pthread_cond_broadcast(&c->out.socket_cond); + c->out.return_socket = 0; + c->out.socket_closure = 0; + _xcb_in_replies_done(c); +} + +static void prepare_socket_request(xcb_connection_t *c) +{ + /* We're about to append data to out.queue, so we need to + * atomically test for an external socket owner *and* some other + * thread currently writing. + * + * If we have an external socket owner, we have to get the socket back + * before we can use it again. + * + * If some other thread is writing to the socket, we assume it's + * writing from out.queue, and so we can't stick data there. + * + * We satisfy this condition by first calling get_socket_back + * (which may drop the lock, but will return when XCB owns the + * socket again) and then checking for another writing thread and + * escaping the loop if we're ready to go. + */ + for (;;) { + if(c->has_error) + return; + get_socket_back(c); + if (!c->out.writing) + break; + pthread_cond_wait(&c->out.cond, &c->iolock); } - return ret; } /* Public interface */ -CARD32 XCBGetMaximumRequestLength(XCBConnection *c) +void xcb_prefetch_maximum_request_length(xcb_connection_t *c) { + if(c->has_error) + return; pthread_mutex_lock(&c->out.reqlenlock); - if(!c->out.maximum_request_length) + if(c->out.maximum_request_length_tag == LAZY_NONE) { - const XCBQueryExtensionRep *ext; - c->out.maximum_request_length = c->setup->maximum_request_length; - ext = XCBGetExtensionData(c, &XCBBigRequestsId); + const xcb_query_extension_reply_t *ext; + ext = xcb_get_extension_data(c, &xcb_big_requests_id); if(ext && ext->present) { - XCBBigRequestsEnableRep *r = XCBBigRequestsEnableReply(c, XCBBigRequestsEnable(c), 0); - c->out.maximum_request_length = r->maximum_request_length; + c->out.maximum_request_length_tag = LAZY_COOKIE; + c->out.maximum_request_length.cookie = xcb_big_requests_enable(c); + } + else + { + c->out.maximum_request_length_tag = LAZY_FORCED; + c->out.maximum_request_length.value = c->setup->maximum_request_length; + } + } + pthread_mutex_unlock(&c->out.reqlenlock); +} + +uint32_t xcb_get_maximum_request_length(xcb_connection_t *c) +{ + if(c->has_error) + return 0; + xcb_prefetch_maximum_request_length(c); + pthread_mutex_lock(&c->out.reqlenlock); + if(c->out.maximum_request_length_tag == LAZY_COOKIE) + { + xcb_big_requests_enable_reply_t *r = xcb_big_requests_enable_reply(c, c->out.maximum_request_length.cookie, 0); + c->out.maximum_request_length_tag = LAZY_FORCED; + if(r) + { + c->out.maximum_request_length.value = r->maximum_request_length; free(r); } + else + c->out.maximum_request_length.value = c->setup->maximum_request_length; } pthread_mutex_unlock(&c->out.reqlenlock); - return c->out.maximum_request_length; + return c->out.maximum_request_length.value; } -unsigned int XCBSendRequest(XCBConnection *c, int flags, struct iovec *vector, const XCBProtocolRequest *req) +unsigned int xcb_send_request(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req) { - unsigned int request; - CARD32 prefix[2]; + uint64_t request; + uint32_t prefix[2]; int veclen = req->count; enum workarounds workaround = WORKAROUND_NONE; + if(c->has_error) + return 0; + assert(c != 0); assert(vector != 0); assert(req->count > 0); @@ -83,21 +194,24 @@ unsigned int XCBSendRequest(XCBConnection *c, int flags, struct iovec *vector, c if(!(flags & XCB_REQUEST_RAW)) { static const char pad[3]; - int i; - CARD16 shortlen = 0; + unsigned int i; + uint16_t shortlen = 0; size_t longlen = 0; assert(vector[0].iov_len >= 4); /* set the major opcode, and the minor opcode for extensions */ if(req->ext) { - const XCBQueryExtensionRep *extension = XCBGetExtensionData(c, req->ext); + const xcb_query_extension_reply_t *extension = xcb_get_extension_data(c, req->ext); if(!(extension && extension->present)) + { + _xcb_conn_shutdown(c, XCB_CONN_CLOSED_EXT_NOTSUPPORTED); return 0; - ((CARD8 *) vector[0].iov_base)[0] = extension->major_opcode; - ((CARD8 *) vector[0].iov_base)[1] = req->opcode; + } + ((uint8_t *) vector[0].iov_base)[0] = extension->major_opcode; + ((uint8_t *) vector[0].iov_base)[1] = req->opcode; } else - ((CARD8 *) vector[0].iov_base)[0] = req->opcode; + ((uint8_t *) vector[0].iov_base)[0] = req->opcode; /* put together the length field, possibly using BIGREQUESTS */ for(i = 0; i < req->count; ++i) @@ -105,7 +219,7 @@ unsigned int XCBSendRequest(XCBConnection *c, int flags, struct iovec *vector, c longlen += vector[i].iov_len; if(!vector[i].iov_base) { - vector[i].iov_base = (caddr_t) pad; + vector[i].iov_base = (char *) pad; assert(vector[i].iov_len <= sizeof(pad)); } } @@ -118,58 +232,124 @@ unsigned int XCBSendRequest(XCBConnection *c, int flags, struct iovec *vector, c shortlen = longlen; longlen = 0; } - else if(longlen > XCBGetMaximumRequestLength(c)) + else if(longlen > xcb_get_maximum_request_length(c)) + { + _xcb_conn_shutdown(c, XCB_CONN_CLOSED_REQ_LEN_EXCEED); return 0; /* server can't take this; maybe need BIGREQUESTS? */ + } /* set the length field. */ - ((CARD16 *) vector[0].iov_base)[1] = shortlen; + ((uint16_t *) vector[0].iov_base)[1] = shortlen; if(!shortlen) { + prefix[0] = ((uint32_t *) vector[0].iov_base)[0]; + prefix[1] = ++longlen; + vector[0].iov_base = (uint32_t *) vector[0].iov_base + 1; + vector[0].iov_len -= sizeof(uint32_t); --vector, ++veclen; vector[0].iov_base = prefix; vector[0].iov_len = sizeof(prefix); - prefix[0] = ((CARD32 *) vector[0].iov_base)[0]; - prefix[1] = ++longlen; - vector[1].iov_base = ((char *) vector[1].iov_base) + sizeof(CARD32); - vector[1].iov_len -= sizeof(CARD32); } } flags &= ~XCB_REQUEST_RAW; /* do we need to work around the X server bug described in glx.xml? */ - if(req->ext && !req->isvoid && strcmp(req->ext->name, "GLX") && - ((req->opcode == 17 && ((CARD32 *) vector[0].iov_base)[0] == 0x10004) || + /* XXX: GetFBConfigs won't use BIG-REQUESTS in any sane + * configuration, but that should be handled here anyway. */ + if(req->ext && !req->isvoid && !strcmp(req->ext->name, "GLX") && + ((req->opcode == 17 && ((uint32_t *) vector[0].iov_base)[1] == 0x10004) || req->opcode == 21)) workaround = WORKAROUND_GLX_GET_FB_CONFIGS_BUG; /* get a sequence number and arrange for delivery. */ pthread_mutex_lock(&c->iolock); - if(req->isvoid && !force_sequence_wrap(c)) + + prepare_socket_request(c); + + /* send GetInputFocus (sync_req) when 64k-2 requests have been sent without + * a reply. + * Also send sync_req (could use NoOp) at 32-bit wrap to avoid having + * applications see sequence 0 as that is used to indicate + * an error in sending the request + */ + + while ((req->isvoid && c->out.request == c->in.request_expected + (1 << 16) - 2) || + (unsigned int) (c->out.request + 1) == 0) { - pthread_mutex_unlock(&c->iolock); - return 0; + send_sync(c); + prepare_socket_request(c); } - /* wait for other writing threads to get out of my way. */ - while(c->out.writing) - pthread_cond_wait(&c->out.cond, &c->iolock); + send_request(c, req->isvoid, workaround, flags, vector, veclen); + request = c->has_error ? 0 : c->out.request; + pthread_mutex_unlock(&c->iolock); + return request; +} - request = ++c->out.request; - assert(request != 0); +void +xcb_send_fd(xcb_connection_t *c, int fd) +{ +#if HAVE_SENDMSG + if (c->has_error) + return; + pthread_mutex_lock(&c->iolock); + while (c->out.out_fd.nfd == XCB_MAX_PASS_FD) { + _xcb_out_flush_to(c, c->out.request); + if (c->has_error) + break; + } + if (!c->has_error) + c->out.out_fd.fd[c->out.out_fd.nfd++] = fd; + pthread_mutex_unlock(&c->iolock); +#endif +} - _xcb_in_expect_reply(c, request, workaround, flags); +int xcb_take_socket(xcb_connection_t *c, void (*return_socket)(void *closure), void *closure, int flags, uint64_t *sent) +{ + int ret; + if(c->has_error) + return 0; + pthread_mutex_lock(&c->iolock); + get_socket_back(c); + + /* _xcb_out_flush may drop the iolock allowing other threads to + * write requests, so keep flushing until we're done + */ + do + ret = _xcb_out_flush_to(c, c->out.request); + while (ret && c->out.request != c->out.request_written); + if(ret) + { + c->out.return_socket = return_socket; + c->out.socket_closure = closure; + if(flags) + _xcb_in_expect_reply(c, c->out.request, WORKAROUND_EXTERNAL_SOCKET_OWNER, flags); + assert(c->out.request == c->out.request_written); + *sent = c->out.request; + } + pthread_mutex_unlock(&c->iolock); + return ret; +} - if(!_xcb_out_write_block(c, vector, veclen)) - request = 0; +int xcb_writev(xcb_connection_t *c, struct iovec *vector, int count, uint64_t requests) +{ + int ret; + if(c->has_error) + return 0; + pthread_mutex_lock(&c->iolock); + c->out.request += requests; + ret = _xcb_out_send(c, vector, count); pthread_mutex_unlock(&c->iolock); - return request; + return ret; } -int XCBFlush(XCBConnection *c) +int xcb_flush(xcb_connection_t *c) { int ret; + if(c->has_error) + return 0; pthread_mutex_lock(&c->iolock); - ret = _xcb_out_flush(c); + ret = _xcb_out_flush_to(c, c->out.request); pthread_mutex_unlock(&c->iolock); return ret; } @@ -178,20 +358,24 @@ int XCBFlush(XCBConnection *c) int _xcb_out_init(_xcb_out *out) { + if(pthread_cond_init(&out->socket_cond, 0)) + return 0; + out->return_socket = 0; + out->socket_closure = 0; + out->socket_moving = 0; + if(pthread_cond_init(&out->cond, 0)) return 0; out->writing = 0; out->queue_len = 0; - out->vec = 0; - out->vec_len = 0; out->request = 0; out->request_written = 0; if(pthread_mutex_init(&out->reqlenlock, 0)) return 0; - out->maximum_request_length = 0; + out->maximum_request_length_tag = LAZY_NONE; return 1; } @@ -202,74 +386,38 @@ void _xcb_out_destroy(_xcb_out *out) pthread_mutex_destroy(&out->reqlenlock); } -/* precondition: there must be something for us to write. */ -int _xcb_out_write(XCBConnection *c) +int _xcb_out_send(xcb_connection_t *c, struct iovec *vector, int count) { - int n; - assert(!c->out.queue_len); - n = writev(c->fd, c->out.vec, c->out.vec_len); - if(n < 0 && errno == EAGAIN) - return 1; - if(n <= 0) - return 0; - - for(; c->out.vec_len; --c->out.vec_len, ++c->out.vec) - { - int cur = c->out.vec->iov_len; - if(cur > n) - cur = n; - c->out.vec->iov_len -= cur; - c->out.vec->iov_base = (char *) c->out.vec->iov_base + cur; - n -= cur; - if(c->out.vec->iov_len) - break; - } - if(!c->out.vec_len) - c->out.vec = 0; - assert(n == 0); - return 1; + int ret = 1; + while(ret && count) + ret = _xcb_conn_wait(c, &c->out.cond, &vector, &count); + c->out.request_written = c->out.request; + pthread_cond_broadcast(&c->out.cond); + _xcb_in_wake_up_next_reader(c); + return ret; } -int _xcb_out_write_block(XCBConnection *c, struct iovec *vector, size_t count) +void _xcb_out_send_sync(xcb_connection_t *c) { - assert(!c->out.vec && !c->out.vec_len); - while(count && c->out.queue_len + vector[0].iov_len < sizeof(c->out.queue)) - { - memcpy(c->out.queue + c->out.queue_len, vector[0].iov_base, vector[0].iov_len); - c->out.queue_len += vector[0].iov_len; - vector[0].iov_base = (char *) vector[0].iov_base + vector[0].iov_len; - vector[0].iov_len = 0; - ++vector, --count; - } - if(!count) - return 1; - - --vector, ++count; - vector[0].iov_base = c->out.queue; - vector[0].iov_len = c->out.queue_len; - c->out.queue_len = 0; - - c->out.vec_len = count; - c->out.vec = vector; - return _xcb_out_flush(c); + prepare_socket_request(c); + send_sync(c); } -int _xcb_out_flush(XCBConnection *c) +int _xcb_out_flush_to(xcb_connection_t *c, uint64_t request) { - int ret = 1; - struct iovec vec; + assert(XCB_SEQUENCE_COMPARE(request, <=, c->out.request)); + if(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request)) + return 1; if(c->out.queue_len) { - assert(!c->out.vec && !c->out.vec_len); + struct iovec vec; vec.iov_base = c->out.queue; vec.iov_len = c->out.queue_len; - c->out.vec = &vec; - c->out.vec_len = 1; c->out.queue_len = 0; + return _xcb_out_send(c, &vec, 1); } - while(ret && c->out.vec_len) - ret = _xcb_conn_wait(c, /*should_write*/ 1, &c->out.cond); - c->out.request_written = c->out.request; - pthread_cond_broadcast(&c->out.cond); - return ret; + while(c->out.writing) + pthread_cond_wait(&c->out.cond, &c->iolock); + assert(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request)); + return 1; }