xcb_out.c revision 709d36bb
1/* Copyright (C) 2001-2004 Bart Massey and Jamey Sharp. 2 * 3 * Permission is hereby granted, free of charge, to any person obtaining a 4 * copy of this software and associated documentation files (the "Software"), 5 * to deal in the Software without restriction, including without limitation 6 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 7 * and/or sell copies of the Software, and to permit persons to whom the 8 * Software is furnished to do so, subject to the following conditions: 9 * 10 * The above copyright notice and this permission notice shall be included in 11 * all copies or substantial portions of the Software. 12 * 13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 16 * AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 17 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 18 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 19 * 20 * Except as contained in this notice, the names of the authors or their 21 * institutions shall not be used in advertising or otherwise to promote the 22 * sale, use or other dealings in this Software without prior written 23 * authorization from the authors. 24 */ 25 26/* Stuff that sends stuff to the server. */ 27 28#ifdef HAVE_CONFIG_H 29#include "config.h" 30#endif 31 32#include <assert.h> 33#include <stdlib.h> 34#include <unistd.h> 35#include <string.h> 36 37#include "xcb.h" 38#include "xcbext.h" 39#include "xcbint.h" 40#include "bigreq.h" 41 42static inline void send_request(xcb_connection_t *c, int isvoid, enum workarounds workaround, int flags, struct iovec *vector, int count) 43{ 44 if(c->has_error) 45 return; 46 47 ++c->out.request; 48 if(!isvoid) 49 c->in.request_expected = c->out.request; 50 if(workaround != WORKAROUND_NONE || flags != 0) 51 _xcb_in_expect_reply(c, c->out.request, workaround, flags); 52 53 while(count && c->out.queue_len + vector[0].iov_len <= sizeof(c->out.queue)) 54 { 55 memcpy(c->out.queue + c->out.queue_len, vector[0].iov_base, vector[0].iov_len); 56 c->out.queue_len += vector[0].iov_len; 57 vector[0].iov_base = (char *) vector[0].iov_base + vector[0].iov_len; 58 vector[0].iov_len = 0; 59 ++vector, --count; 60 } 61 if(!count) 62 return; 63 64 --vector, ++count; 65 vector[0].iov_base = c->out.queue; 66 vector[0].iov_len = c->out.queue_len; 67 c->out.queue_len = 0; 68 _xcb_out_send(c, vector, count); 69} 70 71static void send_sync(xcb_connection_t *c) 72{ 73 static const union { 74 struct { 75 uint8_t major; 76 uint8_t pad; 77 uint16_t len; 78 } fields; 79 uint32_t packet; 80 } sync_req = { { /* GetInputFocus */ 43, 0, 1 } }; 81 struct iovec vector[2]; 82 vector[1].iov_base = (char *) &sync_req; 83 vector[1].iov_len = sizeof(sync_req); 84 send_request(c, 0, WORKAROUND_NONE, XCB_REQUEST_DISCARD_REPLY, vector + 1, 1); 85} 86 87static void get_socket_back(xcb_connection_t *c) 88{ 89 while(c->out.return_socket && c->out.socket_moving) 90 pthread_cond_wait(&c->out.socket_cond, &c->iolock); 91 if(!c->out.return_socket) 92 return; 93 94 c->out.socket_moving = 1; 95 pthread_mutex_unlock(&c->iolock); 96 c->out.return_socket(c->out.socket_closure); 97 pthread_mutex_lock(&c->iolock); 98 c->out.socket_moving = 0; 99 100 pthread_cond_broadcast(&c->out.socket_cond); 101 c->out.return_socket = 0; 102 c->out.socket_closure = 0; 103 _xcb_in_replies_done(c); 104} 105 106static void prepare_socket_request(xcb_connection_t *c) 107{ 108 /* We're about to append data to out.queue, so we need to 109 * atomically test for an external socket owner *and* some other 110 * thread currently writing. 111 * 112 * If we have an external socket owner, we have to get the socket back 113 * before we can use it again. 114 * 115 * If some other thread is writing to the socket, we assume it's 116 * writing from out.queue, and so we can't stick data there. 117 * 118 * We satisfy this condition by first calling get_socket_back 119 * (which may drop the lock, but will return when XCB owns the 120 * socket again) and then checking for another writing thread and 121 * escaping the loop if we're ready to go. 122 */ 123 for (;;) { 124 if(c->has_error) 125 return; 126 get_socket_back(c); 127 if (!c->out.writing) 128 break; 129 pthread_cond_wait(&c->out.cond, &c->iolock); 130 } 131} 132 133/* Public interface */ 134 135void xcb_prefetch_maximum_request_length(xcb_connection_t *c) 136{ 137 if(c->has_error) 138 return; 139 pthread_mutex_lock(&c->out.reqlenlock); 140 if(c->out.maximum_request_length_tag == LAZY_NONE) 141 { 142 const xcb_query_extension_reply_t *ext; 143 ext = xcb_get_extension_data(c, &xcb_big_requests_id); 144 if(ext && ext->present) 145 { 146 c->out.maximum_request_length_tag = LAZY_COOKIE; 147 c->out.maximum_request_length.cookie = xcb_big_requests_enable(c); 148 } 149 else 150 { 151 c->out.maximum_request_length_tag = LAZY_FORCED; 152 c->out.maximum_request_length.value = c->setup->maximum_request_length; 153 } 154 } 155 pthread_mutex_unlock(&c->out.reqlenlock); 156} 157 158uint32_t xcb_get_maximum_request_length(xcb_connection_t *c) 159{ 160 if(c->has_error) 161 return 0; 162 xcb_prefetch_maximum_request_length(c); 163 pthread_mutex_lock(&c->out.reqlenlock); 164 if(c->out.maximum_request_length_tag == LAZY_COOKIE) 165 { 166 xcb_big_requests_enable_reply_t *r = xcb_big_requests_enable_reply(c, c->out.maximum_request_length.cookie, 0); 167 c->out.maximum_request_length_tag = LAZY_FORCED; 168 if(r) 169 { 170 c->out.maximum_request_length.value = r->maximum_request_length; 171 free(r); 172 } 173 else 174 c->out.maximum_request_length.value = c->setup->maximum_request_length; 175 } 176 pthread_mutex_unlock(&c->out.reqlenlock); 177 return c->out.maximum_request_length.value; 178} 179 180uint64_t xcb_send_request64(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req) 181{ 182 uint64_t request; 183 uint32_t prefix[2]; 184 int veclen = req->count; 185 enum workarounds workaround = WORKAROUND_NONE; 186 187 if(c->has_error) 188 return 0; 189 190 assert(c != 0); 191 assert(vector != 0); 192 assert(req->count > 0); 193 194 if(!(flags & XCB_REQUEST_RAW)) 195 { 196 static const char pad[3]; 197 unsigned int i; 198 uint16_t shortlen = 0; 199 size_t longlen = 0; 200 assert(vector[0].iov_len >= 4); 201 /* set the major opcode, and the minor opcode for extensions */ 202 if(req->ext) 203 { 204 const xcb_query_extension_reply_t *extension = xcb_get_extension_data(c, req->ext); 205 if(!(extension && extension->present)) 206 { 207 _xcb_conn_shutdown(c, XCB_CONN_CLOSED_EXT_NOTSUPPORTED); 208 return 0; 209 } 210 ((uint8_t *) vector[0].iov_base)[0] = extension->major_opcode; 211 ((uint8_t *) vector[0].iov_base)[1] = req->opcode; 212 } 213 else 214 ((uint8_t *) vector[0].iov_base)[0] = req->opcode; 215 216 /* put together the length field, possibly using BIGREQUESTS */ 217 for(i = 0; i < req->count; ++i) 218 { 219 longlen += vector[i].iov_len; 220 if(!vector[i].iov_base) 221 { 222 vector[i].iov_base = (char *) pad; 223 assert(vector[i].iov_len <= sizeof(pad)); 224 } 225 } 226 assert((longlen & 3) == 0); 227 longlen >>= 2; 228 229 if(longlen <= c->setup->maximum_request_length) 230 { 231 /* we don't need BIGREQUESTS. */ 232 shortlen = longlen; 233 longlen = 0; 234 } 235 else if(longlen > xcb_get_maximum_request_length(c)) 236 { 237 _xcb_conn_shutdown(c, XCB_CONN_CLOSED_REQ_LEN_EXCEED); 238 return 0; /* server can't take this; maybe need BIGREQUESTS? */ 239 } 240 241 /* set the length field. */ 242 ((uint16_t *) vector[0].iov_base)[1] = shortlen; 243 if(!shortlen) 244 { 245 prefix[0] = ((uint32_t *) vector[0].iov_base)[0]; 246 prefix[1] = ++longlen; 247 vector[0].iov_base = (uint32_t *) vector[0].iov_base + 1; 248 vector[0].iov_len -= sizeof(uint32_t); 249 --vector, ++veclen; 250 vector[0].iov_base = prefix; 251 vector[0].iov_len = sizeof(prefix); 252 } 253 } 254 flags &= ~XCB_REQUEST_RAW; 255 256 /* do we need to work around the X server bug described in glx.xml? */ 257 /* XXX: GetFBConfigs won't use BIG-REQUESTS in any sane 258 * configuration, but that should be handled here anyway. */ 259 if(req->ext && !req->isvoid && !strcmp(req->ext->name, "GLX") && 260 ((req->opcode == 17 && ((uint32_t *) vector[0].iov_base)[1] == 0x10004) || 261 req->opcode == 21)) 262 workaround = WORKAROUND_GLX_GET_FB_CONFIGS_BUG; 263 264 /* get a sequence number and arrange for delivery. */ 265 pthread_mutex_lock(&c->iolock); 266 267 prepare_socket_request(c); 268 269 /* send GetInputFocus (sync_req) when 64k-2 requests have been sent without 270 * a reply. 271 * Also send sync_req (could use NoOp) at 32-bit wrap to avoid having 272 * applications see sequence 0 as that is used to indicate 273 * an error in sending the request 274 */ 275 276 while ((req->isvoid && c->out.request == c->in.request_expected + (1 << 16) - 2) || 277 (unsigned int) (c->out.request + 1) == 0) 278 { 279 send_sync(c); 280 prepare_socket_request(c); 281 } 282 283 send_request(c, req->isvoid, workaround, flags, vector, veclen); 284 request = c->has_error ? 0 : c->out.request; 285 pthread_mutex_unlock(&c->iolock); 286 return request; 287} 288 289/* request number are actually uint64_t internally but keep API compat with unsigned int */ 290unsigned int xcb_send_request(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req) 291{ 292 return xcb_send_request64(c, flags, vector, req); 293} 294 295void 296xcb_send_fd(xcb_connection_t *c, int fd) 297{ 298#if HAVE_SENDMSG 299 if (c->has_error) 300 return; 301 pthread_mutex_lock(&c->iolock); 302 while (c->out.out_fd.nfd == XCB_MAX_PASS_FD) { 303 _xcb_out_flush_to(c, c->out.request); 304 if (c->has_error) 305 break; 306 } 307 if (!c->has_error) 308 c->out.out_fd.fd[c->out.out_fd.nfd++] = fd; 309 pthread_mutex_unlock(&c->iolock); 310#endif 311} 312 313int xcb_take_socket(xcb_connection_t *c, void (*return_socket)(void *closure), void *closure, int flags, uint64_t *sent) 314{ 315 int ret; 316 if(c->has_error) 317 return 0; 318 pthread_mutex_lock(&c->iolock); 319 get_socket_back(c); 320 321 /* _xcb_out_flush may drop the iolock allowing other threads to 322 * write requests, so keep flushing until we're done 323 */ 324 do 325 ret = _xcb_out_flush_to(c, c->out.request); 326 while (ret && c->out.request != c->out.request_written); 327 if(ret) 328 { 329 c->out.return_socket = return_socket; 330 c->out.socket_closure = closure; 331 if(flags) 332 _xcb_in_expect_reply(c, c->out.request, WORKAROUND_EXTERNAL_SOCKET_OWNER, flags); 333 assert(c->out.request == c->out.request_written); 334 *sent = c->out.request; 335 } 336 pthread_mutex_unlock(&c->iolock); 337 return ret; 338} 339 340int xcb_writev(xcb_connection_t *c, struct iovec *vector, int count, uint64_t requests) 341{ 342 int ret; 343 if(c->has_error) 344 return 0; 345 pthread_mutex_lock(&c->iolock); 346 c->out.request += requests; 347 ret = _xcb_out_send(c, vector, count); 348 pthread_mutex_unlock(&c->iolock); 349 return ret; 350} 351 352int xcb_flush(xcb_connection_t *c) 353{ 354 int ret; 355 if(c->has_error) 356 return 0; 357 pthread_mutex_lock(&c->iolock); 358 ret = _xcb_out_flush_to(c, c->out.request); 359 pthread_mutex_unlock(&c->iolock); 360 return ret; 361} 362 363/* Private interface */ 364 365int _xcb_out_init(_xcb_out *out) 366{ 367 if(pthread_cond_init(&out->socket_cond, 0)) 368 return 0; 369 out->return_socket = 0; 370 out->socket_closure = 0; 371 out->socket_moving = 0; 372 373 if(pthread_cond_init(&out->cond, 0)) 374 return 0; 375 out->writing = 0; 376 377 out->queue_len = 0; 378 379 out->request = 0; 380 out->request_written = 0; 381 382 if(pthread_mutex_init(&out->reqlenlock, 0)) 383 return 0; 384 out->maximum_request_length_tag = LAZY_NONE; 385 386 return 1; 387} 388 389void _xcb_out_destroy(_xcb_out *out) 390{ 391 pthread_cond_destroy(&out->cond); 392 pthread_mutex_destroy(&out->reqlenlock); 393} 394 395int _xcb_out_send(xcb_connection_t *c, struct iovec *vector, int count) 396{ 397 int ret = 1; 398 while(ret && count) 399 ret = _xcb_conn_wait(c, &c->out.cond, &vector, &count); 400 c->out.request_written = c->out.request; 401 pthread_cond_broadcast(&c->out.cond); 402 _xcb_in_wake_up_next_reader(c); 403 return ret; 404} 405 406void _xcb_out_send_sync(xcb_connection_t *c) 407{ 408 prepare_socket_request(c); 409 send_sync(c); 410} 411 412int _xcb_out_flush_to(xcb_connection_t *c, uint64_t request) 413{ 414 assert(XCB_SEQUENCE_COMPARE(request, <=, c->out.request)); 415 if(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request)) 416 return 1; 417 if(c->out.queue_len) 418 { 419 struct iovec vec; 420 vec.iov_base = c->out.queue; 421 vec.iov_len = c->out.queue_len; 422 c->out.queue_len = 0; 423 return _xcb_out_send(c, &vec, 1); 424 } 425 while(c->out.writing) 426 pthread_cond_wait(&c->out.cond, &c->iolock); 427 assert(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request)); 428 return 1; 429} 430