xcb_out.c revision fe12f63c
1/* Copyright (C) 2001-2004 Bart Massey and Jamey Sharp.
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a
4 * copy of this software and associated documentation files (the "Software"),
5 * to deal in the Software without restriction, including without limitation
6 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
7 * and/or sell copies of the Software, and to permit persons to whom the
8 * Software is furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
17 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
18 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
19 *
20 * Except as contained in this notice, the names of the authors or their
21 * institutions shall not be used in advertising or otherwise to promote the
22 * sale, use or other dealings in this Software without prior written
23 * authorization from the authors.
24 */
25
26/* Stuff that sends stuff to the server. */
27
28#ifdef HAVE_CONFIG_H
29#include "config.h"
30#endif
31
32#include <assert.h>
33#include <stdlib.h>
34#include <unistd.h>
35#include <string.h>
36
37#include "xcb.h"
38#include "xcbext.h"
39#include "xcbint.h"
40#include "bigreq.h"
41
42static inline void send_request(xcb_connection_t *c, int isvoid, enum workarounds workaround, int flags, struct iovec *vector, int count)
43{
44    if(c->has_error)
45        return;
46
47    ++c->out.request;
48    if(!isvoid)
49        c->in.request_expected = c->out.request;
50    if(workaround != WORKAROUND_NONE || flags != 0)
51        _xcb_in_expect_reply(c, c->out.request, workaround, flags);
52
53    while(count && c->out.queue_len + vector[0].iov_len <= sizeof(c->out.queue))
54    {
55        memcpy(c->out.queue + c->out.queue_len, vector[0].iov_base, vector[0].iov_len);
56        c->out.queue_len += vector[0].iov_len;
57        vector[0].iov_base = (char *) vector[0].iov_base + vector[0].iov_len;
58        vector[0].iov_len = 0;
59        ++vector, --count;
60    }
61    if(!count)
62        return;
63
64    --vector, ++count;
65    vector[0].iov_base = c->out.queue;
66    vector[0].iov_len = c->out.queue_len;
67    c->out.queue_len = 0;
68    _xcb_out_send(c, vector, count);
69}
70
71static void send_sync(xcb_connection_t *c)
72{
73    static const union {
74        struct {
75            uint8_t major;
76            uint8_t pad;
77            uint16_t len;
78        } fields;
79        uint32_t packet;
80    } sync_req = { { /* GetInputFocus */ 43, 0, 1 } };
81    struct iovec vector[2];
82    vector[1].iov_base = (char *) &sync_req;
83    vector[1].iov_len = sizeof(sync_req);
84    send_request(c, 0, WORKAROUND_NONE, XCB_REQUEST_DISCARD_REPLY, vector + 1, 1);
85}
86
87static void get_socket_back(xcb_connection_t *c)
88{
89    while(c->out.return_socket && c->out.socket_moving)
90        pthread_cond_wait(&c->out.socket_cond, &c->iolock);
91    if(!c->out.return_socket)
92        return;
93
94    c->out.socket_moving = 1;
95    pthread_mutex_unlock(&c->iolock);
96    c->out.return_socket(c->out.socket_closure);
97    pthread_mutex_lock(&c->iolock);
98    c->out.socket_moving = 0;
99
100    pthread_cond_broadcast(&c->out.socket_cond);
101    c->out.return_socket = 0;
102    c->out.socket_closure = 0;
103    _xcb_in_replies_done(c);
104}
105
106static void prepare_socket_request(xcb_connection_t *c)
107{
108    /* We're about to append data to out.queue, so we need to
109     * atomically test for an external socket owner *and* some other
110     * thread currently writing.
111     *
112     * If we have an external socket owner, we have to get the socket back
113     * before we can use it again.
114     *
115     * If some other thread is writing to the socket, we assume it's
116     * writing from out.queue, and so we can't stick data there.
117     *
118     * We satisfy this condition by first calling get_socket_back
119     * (which may drop the lock, but will return when XCB owns the
120     * socket again) and then checking for another writing thread and
121     * escaping the loop if we're ready to go.
122     */
123    for (;;) {
124        if(c->has_error)
125            return;
126        get_socket_back(c);
127        if (!c->out.writing)
128            break;
129        pthread_cond_wait(&c->out.cond, &c->iolock);
130    }
131}
132
133/* Public interface */
134
135void xcb_prefetch_maximum_request_length(xcb_connection_t *c)
136{
137    if(c->has_error)
138        return;
139    pthread_mutex_lock(&c->out.reqlenlock);
140    if(c->out.maximum_request_length_tag == LAZY_NONE)
141    {
142        const xcb_query_extension_reply_t *ext;
143        ext = xcb_get_extension_data(c, &xcb_big_requests_id);
144        if(ext && ext->present)
145        {
146            c->out.maximum_request_length_tag = LAZY_COOKIE;
147            c->out.maximum_request_length.cookie = xcb_big_requests_enable(c);
148        }
149        else
150        {
151            c->out.maximum_request_length_tag = LAZY_FORCED;
152            c->out.maximum_request_length.value = c->setup->maximum_request_length;
153        }
154    }
155    pthread_mutex_unlock(&c->out.reqlenlock);
156}
157
158uint32_t xcb_get_maximum_request_length(xcb_connection_t *c)
159{
160    if(c->has_error)
161        return 0;
162    xcb_prefetch_maximum_request_length(c);
163    pthread_mutex_lock(&c->out.reqlenlock);
164    if(c->out.maximum_request_length_tag == LAZY_COOKIE)
165    {
166        xcb_big_requests_enable_reply_t *r = xcb_big_requests_enable_reply(c, c->out.maximum_request_length.cookie, 0);
167        c->out.maximum_request_length_tag = LAZY_FORCED;
168        if(r)
169        {
170            c->out.maximum_request_length.value = r->maximum_request_length;
171            free(r);
172        }
173        else
174            c->out.maximum_request_length.value = c->setup->maximum_request_length;
175    }
176    pthread_mutex_unlock(&c->out.reqlenlock);
177    return c->out.maximum_request_length.value;
178}
179
180static void close_fds(int *fds, unsigned int num_fds)
181{
182    for (unsigned int index = 0; index < num_fds; index++)
183        close(fds[index]);
184}
185
186static void send_fds(xcb_connection_t *c, int *fds, unsigned int num_fds)
187{
188#if HAVE_SENDMSG
189    /* Calling _xcb_out_flush_to() can drop the iolock and wait on a condition
190     * variable if another thread is currently writing (c->out.writing > 0).
191     * This call waits for writers to be done and thus _xcb_out_flush_to() will
192     * do the work itself (in which case we are a writer and
193     * prepare_socket_request() will wait for us to be done if another threads
194     * tries to send fds, too). Thanks to this, we can atomically write out FDs.
195     */
196    prepare_socket_request(c);
197
198    while (num_fds > 0) {
199        while (c->out.out_fd.nfd == XCB_MAX_PASS_FD && !c->has_error) {
200            /* XXX: if c->out.writing > 0, this releases the iolock and
201             * potentially allows other threads to interfere with their own fds.
202             */
203            _xcb_out_flush_to(c, c->out.request);
204
205            if (c->out.out_fd.nfd == XCB_MAX_PASS_FD) {
206                /* We need some request to send FDs with */
207                _xcb_out_send_sync(c);
208            }
209        }
210        if (c->has_error)
211            break;
212
213        c->out.out_fd.fd[c->out.out_fd.nfd++] = fds[0];
214        fds++;
215        num_fds--;
216    }
217#endif
218    close_fds(fds, num_fds);
219}
220
221uint64_t xcb_send_request_with_fds64(xcb_connection_t *c, int flags, struct iovec *vector,
222                const xcb_protocol_request_t *req, unsigned int num_fds, int *fds)
223{
224    uint64_t request;
225    uint32_t prefix[2];
226    int veclen = req->count;
227    enum workarounds workaround = WORKAROUND_NONE;
228
229    if(c->has_error) {
230        close_fds(fds, num_fds);
231        return 0;
232    }
233
234    assert(c != 0);
235    assert(vector != 0);
236    assert(req->count > 0);
237
238    if(!(flags & XCB_REQUEST_RAW))
239    {
240        static const char pad[3];
241        unsigned int i;
242        uint16_t shortlen = 0;
243        size_t longlen = 0;
244        assert(vector[0].iov_len >= 4);
245        /* set the major opcode, and the minor opcode for extensions */
246        if(req->ext)
247        {
248            const xcb_query_extension_reply_t *extension = xcb_get_extension_data(c, req->ext);
249            if(!(extension && extension->present))
250            {
251                close_fds(fds, num_fds);
252                _xcb_conn_shutdown(c, XCB_CONN_CLOSED_EXT_NOTSUPPORTED);
253                return 0;
254            }
255            ((uint8_t *) vector[0].iov_base)[0] = extension->major_opcode;
256            ((uint8_t *) vector[0].iov_base)[1] = req->opcode;
257        }
258        else
259            ((uint8_t *) vector[0].iov_base)[0] = req->opcode;
260
261        /* put together the length field, possibly using BIGREQUESTS */
262        for(i = 0; i < req->count; ++i)
263        {
264            longlen += vector[i].iov_len;
265            if(!vector[i].iov_base)
266            {
267                vector[i].iov_base = (char *) pad;
268                assert(vector[i].iov_len <= sizeof(pad));
269            }
270        }
271        assert((longlen & 3) == 0);
272        longlen >>= 2;
273
274        if(longlen <= c->setup->maximum_request_length)
275        {
276            /* we don't need BIGREQUESTS. */
277            shortlen = longlen;
278            longlen = 0;
279        }
280        else if(longlen > xcb_get_maximum_request_length(c))
281        {
282            close_fds(fds, num_fds);
283            _xcb_conn_shutdown(c, XCB_CONN_CLOSED_REQ_LEN_EXCEED);
284            return 0; /* server can't take this; maybe need BIGREQUESTS? */
285        }
286
287        /* set the length field. */
288        ((uint16_t *) vector[0].iov_base)[1] = shortlen;
289        if(!shortlen)
290        {
291            prefix[0] = ((uint32_t *) vector[0].iov_base)[0];
292            prefix[1] = ++longlen;
293            vector[0].iov_base = (uint32_t *) vector[0].iov_base + 1;
294            vector[0].iov_len -= sizeof(uint32_t);
295            --vector, ++veclen;
296            vector[0].iov_base = prefix;
297            vector[0].iov_len = sizeof(prefix);
298        }
299    }
300    flags &= ~XCB_REQUEST_RAW;
301
302    /* do we need to work around the X server bug described in glx.xml? */
303    /* XXX: GetFBConfigs won't use BIG-REQUESTS in any sane
304     * configuration, but that should be handled here anyway. */
305    if(req->ext && !req->isvoid && !strcmp(req->ext->name, "GLX") &&
306            ((req->opcode == 17 && ((uint32_t *) vector[0].iov_base)[1] == 0x10004) ||
307             req->opcode == 21))
308        workaround = WORKAROUND_GLX_GET_FB_CONFIGS_BUG;
309
310    /* get a sequence number and arrange for delivery. */
311    pthread_mutex_lock(&c->iolock);
312
313    /* send FDs before establishing a good request number, because this might
314     * call send_sync(), too
315     */
316    send_fds(c, fds, num_fds);
317
318    prepare_socket_request(c);
319
320    /* send GetInputFocus (sync_req) when 64k-2 requests have been sent without
321     * a reply.
322     * Also send sync_req (could use NoOp) at 32-bit wrap to avoid having
323     * applications see sequence 0 as that is used to indicate
324     * an error in sending the request
325     */
326
327    while ((req->isvoid && c->out.request == c->in.request_expected + (1 << 16) - 2) ||
328           (unsigned int) (c->out.request + 1) == 0)
329    {
330        send_sync(c);
331        prepare_socket_request(c);
332    }
333
334    send_request(c, req->isvoid, workaround, flags, vector, veclen);
335    request = c->has_error ? 0 : c->out.request;
336    pthread_mutex_unlock(&c->iolock);
337    return request;
338}
339
340/* request number are actually uint64_t internally but keep API compat with unsigned int */
341unsigned int xcb_send_request_with_fds(xcb_connection_t *c, int flags, struct iovec *vector,
342        const xcb_protocol_request_t *req, unsigned int num_fds, int *fds)
343{
344    return xcb_send_request_with_fds64(c, flags, vector, req, num_fds, fds);
345}
346
347uint64_t xcb_send_request64(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req)
348{
349    return xcb_send_request_with_fds64(c, flags, vector, req, 0, NULL);
350}
351
352/* request number are actually uint64_t internally but keep API compat with unsigned int */
353unsigned int xcb_send_request(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req)
354{
355    return xcb_send_request64(c, flags, vector, req);
356}
357
358void
359xcb_send_fd(xcb_connection_t *c, int fd)
360{
361    int fds[1] = { fd };
362
363    if (c->has_error) {
364        close(fd);
365        return;
366    }
367    pthread_mutex_lock(&c->iolock);
368    send_fds(c, &fds[0], 1);
369    pthread_mutex_unlock(&c->iolock);
370}
371
372int xcb_take_socket(xcb_connection_t *c, void (*return_socket)(void *closure), void *closure, int flags, uint64_t *sent)
373{
374    int ret;
375    if(c->has_error)
376        return 0;
377    pthread_mutex_lock(&c->iolock);
378    get_socket_back(c);
379
380    /* _xcb_out_flush may drop the iolock allowing other threads to
381     * write requests, so keep flushing until we're done
382     */
383    do
384        ret = _xcb_out_flush_to(c, c->out.request);
385    while (ret && c->out.request != c->out.request_written);
386    if(ret)
387    {
388        c->out.return_socket = return_socket;
389        c->out.socket_closure = closure;
390        if(flags) {
391            /* c->out.request + 1 will be the first request sent by the external
392             * socket owner. If the socket is returned before this request is sent
393             * it will be detected in _xcb_in_replies_done and this pending_reply
394             * will be discarded.
395             */
396            _xcb_in_expect_reply(c, c->out.request + 1, WORKAROUND_EXTERNAL_SOCKET_OWNER, flags);
397        }
398        assert(c->out.request == c->out.request_written);
399        *sent = c->out.request;
400    }
401    pthread_mutex_unlock(&c->iolock);
402    return ret;
403}
404
405int xcb_writev(xcb_connection_t *c, struct iovec *vector, int count, uint64_t requests)
406{
407    int ret;
408    if(c->has_error)
409        return 0;
410    pthread_mutex_lock(&c->iolock);
411    c->out.request += requests;
412    ret = _xcb_out_send(c, vector, count);
413    pthread_mutex_unlock(&c->iolock);
414    return ret;
415}
416
417int xcb_flush(xcb_connection_t *c)
418{
419    int ret;
420    if(c->has_error)
421        return 0;
422    pthread_mutex_lock(&c->iolock);
423    ret = _xcb_out_flush_to(c, c->out.request);
424    pthread_mutex_unlock(&c->iolock);
425    return ret;
426}
427
428/* Private interface */
429
430int _xcb_out_init(_xcb_out *out)
431{
432    if(pthread_cond_init(&out->socket_cond, 0))
433        return 0;
434    out->return_socket = 0;
435    out->socket_closure = 0;
436    out->socket_moving = 0;
437
438    if(pthread_cond_init(&out->cond, 0))
439        return 0;
440    out->writing = 0;
441
442    out->queue_len = 0;
443
444    out->request = 0;
445    out->request_written = 0;
446
447    if(pthread_mutex_init(&out->reqlenlock, 0))
448        return 0;
449    out->maximum_request_length_tag = LAZY_NONE;
450
451    return 1;
452}
453
454void _xcb_out_destroy(_xcb_out *out)
455{
456    pthread_cond_destroy(&out->cond);
457    pthread_mutex_destroy(&out->reqlenlock);
458}
459
460int _xcb_out_send(xcb_connection_t *c, struct iovec *vector, int count)
461{
462    int ret = 1;
463    while(ret && count)
464        ret = _xcb_conn_wait(c, &c->out.cond, &vector, &count);
465    c->out.request_written = c->out.request;
466    pthread_cond_broadcast(&c->out.cond);
467    _xcb_in_wake_up_next_reader(c);
468    return ret;
469}
470
471void _xcb_out_send_sync(xcb_connection_t *c)
472{
473    prepare_socket_request(c);
474    send_sync(c);
475}
476
477int _xcb_out_flush_to(xcb_connection_t *c, uint64_t request)
478{
479    assert(XCB_SEQUENCE_COMPARE(request, <=, c->out.request));
480    if(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request))
481        return 1;
482    if(c->out.queue_len)
483    {
484        struct iovec vec;
485        vec.iov_base = c->out.queue;
486        vec.iov_len = c->out.queue_len;
487        c->out.queue_len = 0;
488        return _xcb_out_send(c, &vec, 1);
489    }
490    while(c->out.writing)
491        pthread_cond_wait(&c->out.cond, &c->iolock);
492    assert(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request));
493    return 1;
494}
495