xcb_out.c revision 709d36bb
1/* Copyright (C) 2001-2004 Bart Massey and Jamey Sharp.
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a
4 * copy of this software and associated documentation files (the "Software"),
5 * to deal in the Software without restriction, including without limitation
6 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
7 * and/or sell copies of the Software, and to permit persons to whom the
8 * Software is furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
17 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
18 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
19 *
20 * Except as contained in this notice, the names of the authors or their
21 * institutions shall not be used in advertising or otherwise to promote the
22 * sale, use or other dealings in this Software without prior written
23 * authorization from the authors.
24 */
25
26/* Stuff that sends stuff to the server. */
27
28#ifdef HAVE_CONFIG_H
29#include "config.h"
30#endif
31
32#include <assert.h>
33#include <stdlib.h>
34#include <unistd.h>
35#include <string.h>
36
37#include "xcb.h"
38#include "xcbext.h"
39#include "xcbint.h"
40#include "bigreq.h"
41
42static inline void send_request(xcb_connection_t *c, int isvoid, enum workarounds workaround, int flags, struct iovec *vector, int count)
43{
44    if(c->has_error)
45        return;
46
47    ++c->out.request;
48    if(!isvoid)
49        c->in.request_expected = c->out.request;
50    if(workaround != WORKAROUND_NONE || flags != 0)
51        _xcb_in_expect_reply(c, c->out.request, workaround, flags);
52
53    while(count && c->out.queue_len + vector[0].iov_len <= sizeof(c->out.queue))
54    {
55        memcpy(c->out.queue + c->out.queue_len, vector[0].iov_base, vector[0].iov_len);
56        c->out.queue_len += vector[0].iov_len;
57        vector[0].iov_base = (char *) vector[0].iov_base + vector[0].iov_len;
58        vector[0].iov_len = 0;
59        ++vector, --count;
60    }
61    if(!count)
62        return;
63
64    --vector, ++count;
65    vector[0].iov_base = c->out.queue;
66    vector[0].iov_len = c->out.queue_len;
67    c->out.queue_len = 0;
68    _xcb_out_send(c, vector, count);
69}
70
71static void send_sync(xcb_connection_t *c)
72{
73    static const union {
74        struct {
75            uint8_t major;
76            uint8_t pad;
77            uint16_t len;
78        } fields;
79        uint32_t packet;
80    } sync_req = { { /* GetInputFocus */ 43, 0, 1 } };
81    struct iovec vector[2];
82    vector[1].iov_base = (char *) &sync_req;
83    vector[1].iov_len = sizeof(sync_req);
84    send_request(c, 0, WORKAROUND_NONE, XCB_REQUEST_DISCARD_REPLY, vector + 1, 1);
85}
86
87static void get_socket_back(xcb_connection_t *c)
88{
89    while(c->out.return_socket && c->out.socket_moving)
90        pthread_cond_wait(&c->out.socket_cond, &c->iolock);
91    if(!c->out.return_socket)
92        return;
93
94    c->out.socket_moving = 1;
95    pthread_mutex_unlock(&c->iolock);
96    c->out.return_socket(c->out.socket_closure);
97    pthread_mutex_lock(&c->iolock);
98    c->out.socket_moving = 0;
99
100    pthread_cond_broadcast(&c->out.socket_cond);
101    c->out.return_socket = 0;
102    c->out.socket_closure = 0;
103    _xcb_in_replies_done(c);
104}
105
106static void prepare_socket_request(xcb_connection_t *c)
107{
108    /* We're about to append data to out.queue, so we need to
109     * atomically test for an external socket owner *and* some other
110     * thread currently writing.
111     *
112     * If we have an external socket owner, we have to get the socket back
113     * before we can use it again.
114     *
115     * If some other thread is writing to the socket, we assume it's
116     * writing from out.queue, and so we can't stick data there.
117     *
118     * We satisfy this condition by first calling get_socket_back
119     * (which may drop the lock, but will return when XCB owns the
120     * socket again) and then checking for another writing thread and
121     * escaping the loop if we're ready to go.
122     */
123    for (;;) {
124        if(c->has_error)
125            return;
126        get_socket_back(c);
127        if (!c->out.writing)
128            break;
129        pthread_cond_wait(&c->out.cond, &c->iolock);
130    }
131}
132
133/* Public interface */
134
135void xcb_prefetch_maximum_request_length(xcb_connection_t *c)
136{
137    if(c->has_error)
138        return;
139    pthread_mutex_lock(&c->out.reqlenlock);
140    if(c->out.maximum_request_length_tag == LAZY_NONE)
141    {
142        const xcb_query_extension_reply_t *ext;
143        ext = xcb_get_extension_data(c, &xcb_big_requests_id);
144        if(ext && ext->present)
145        {
146            c->out.maximum_request_length_tag = LAZY_COOKIE;
147            c->out.maximum_request_length.cookie = xcb_big_requests_enable(c);
148        }
149        else
150        {
151            c->out.maximum_request_length_tag = LAZY_FORCED;
152            c->out.maximum_request_length.value = c->setup->maximum_request_length;
153        }
154    }
155    pthread_mutex_unlock(&c->out.reqlenlock);
156}
157
158uint32_t xcb_get_maximum_request_length(xcb_connection_t *c)
159{
160    if(c->has_error)
161        return 0;
162    xcb_prefetch_maximum_request_length(c);
163    pthread_mutex_lock(&c->out.reqlenlock);
164    if(c->out.maximum_request_length_tag == LAZY_COOKIE)
165    {
166        xcb_big_requests_enable_reply_t *r = xcb_big_requests_enable_reply(c, c->out.maximum_request_length.cookie, 0);
167        c->out.maximum_request_length_tag = LAZY_FORCED;
168        if(r)
169        {
170            c->out.maximum_request_length.value = r->maximum_request_length;
171            free(r);
172        }
173        else
174            c->out.maximum_request_length.value = c->setup->maximum_request_length;
175    }
176    pthread_mutex_unlock(&c->out.reqlenlock);
177    return c->out.maximum_request_length.value;
178}
179
180uint64_t xcb_send_request64(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req)
181{
182    uint64_t request;
183    uint32_t prefix[2];
184    int veclen = req->count;
185    enum workarounds workaround = WORKAROUND_NONE;
186
187    if(c->has_error)
188        return 0;
189
190    assert(c != 0);
191    assert(vector != 0);
192    assert(req->count > 0);
193
194    if(!(flags & XCB_REQUEST_RAW))
195    {
196        static const char pad[3];
197        unsigned int i;
198        uint16_t shortlen = 0;
199        size_t longlen = 0;
200        assert(vector[0].iov_len >= 4);
201        /* set the major opcode, and the minor opcode for extensions */
202        if(req->ext)
203        {
204            const xcb_query_extension_reply_t *extension = xcb_get_extension_data(c, req->ext);
205            if(!(extension && extension->present))
206            {
207                _xcb_conn_shutdown(c, XCB_CONN_CLOSED_EXT_NOTSUPPORTED);
208                return 0;
209            }
210            ((uint8_t *) vector[0].iov_base)[0] = extension->major_opcode;
211            ((uint8_t *) vector[0].iov_base)[1] = req->opcode;
212        }
213        else
214            ((uint8_t *) vector[0].iov_base)[0] = req->opcode;
215
216        /* put together the length field, possibly using BIGREQUESTS */
217        for(i = 0; i < req->count; ++i)
218        {
219            longlen += vector[i].iov_len;
220            if(!vector[i].iov_base)
221            {
222                vector[i].iov_base = (char *) pad;
223                assert(vector[i].iov_len <= sizeof(pad));
224            }
225        }
226        assert((longlen & 3) == 0);
227        longlen >>= 2;
228
229        if(longlen <= c->setup->maximum_request_length)
230        {
231            /* we don't need BIGREQUESTS. */
232            shortlen = longlen;
233            longlen = 0;
234        }
235        else if(longlen > xcb_get_maximum_request_length(c))
236        {
237            _xcb_conn_shutdown(c, XCB_CONN_CLOSED_REQ_LEN_EXCEED);
238            return 0; /* server can't take this; maybe need BIGREQUESTS? */
239        }
240
241        /* set the length field. */
242        ((uint16_t *) vector[0].iov_base)[1] = shortlen;
243        if(!shortlen)
244        {
245            prefix[0] = ((uint32_t *) vector[0].iov_base)[0];
246            prefix[1] = ++longlen;
247            vector[0].iov_base = (uint32_t *) vector[0].iov_base + 1;
248            vector[0].iov_len -= sizeof(uint32_t);
249            --vector, ++veclen;
250            vector[0].iov_base = prefix;
251            vector[0].iov_len = sizeof(prefix);
252        }
253    }
254    flags &= ~XCB_REQUEST_RAW;
255
256    /* do we need to work around the X server bug described in glx.xml? */
257    /* XXX: GetFBConfigs won't use BIG-REQUESTS in any sane
258     * configuration, but that should be handled here anyway. */
259    if(req->ext && !req->isvoid && !strcmp(req->ext->name, "GLX") &&
260            ((req->opcode == 17 && ((uint32_t *) vector[0].iov_base)[1] == 0x10004) ||
261             req->opcode == 21))
262        workaround = WORKAROUND_GLX_GET_FB_CONFIGS_BUG;
263
264    /* get a sequence number and arrange for delivery. */
265    pthread_mutex_lock(&c->iolock);
266
267    prepare_socket_request(c);
268
269    /* send GetInputFocus (sync_req) when 64k-2 requests have been sent without
270     * a reply.
271     * Also send sync_req (could use NoOp) at 32-bit wrap to avoid having
272     * applications see sequence 0 as that is used to indicate
273     * an error in sending the request
274     */
275
276    while ((req->isvoid && c->out.request == c->in.request_expected + (1 << 16) - 2) ||
277           (unsigned int) (c->out.request + 1) == 0)
278    {
279        send_sync(c);
280        prepare_socket_request(c);
281    }
282
283    send_request(c, req->isvoid, workaround, flags, vector, veclen);
284    request = c->has_error ? 0 : c->out.request;
285    pthread_mutex_unlock(&c->iolock);
286    return request;
287}
288
289/* request number are actually uint64_t internally but keep API compat with unsigned int */
290unsigned int xcb_send_request(xcb_connection_t *c, int flags, struct iovec *vector, const xcb_protocol_request_t *req)
291{
292    return xcb_send_request64(c, flags, vector, req);
293}
294
295void
296xcb_send_fd(xcb_connection_t *c, int fd)
297{
298#if HAVE_SENDMSG
299    if (c->has_error)
300        return;
301    pthread_mutex_lock(&c->iolock);
302    while (c->out.out_fd.nfd == XCB_MAX_PASS_FD) {
303        _xcb_out_flush_to(c, c->out.request);
304        if (c->has_error)
305            break;
306    }
307    if (!c->has_error)
308        c->out.out_fd.fd[c->out.out_fd.nfd++] = fd;
309    pthread_mutex_unlock(&c->iolock);
310#endif
311}
312
313int xcb_take_socket(xcb_connection_t *c, void (*return_socket)(void *closure), void *closure, int flags, uint64_t *sent)
314{
315    int ret;
316    if(c->has_error)
317        return 0;
318    pthread_mutex_lock(&c->iolock);
319    get_socket_back(c);
320
321    /* _xcb_out_flush may drop the iolock allowing other threads to
322     * write requests, so keep flushing until we're done
323     */
324    do
325        ret = _xcb_out_flush_to(c, c->out.request);
326    while (ret && c->out.request != c->out.request_written);
327    if(ret)
328    {
329        c->out.return_socket = return_socket;
330        c->out.socket_closure = closure;
331        if(flags)
332            _xcb_in_expect_reply(c, c->out.request, WORKAROUND_EXTERNAL_SOCKET_OWNER, flags);
333        assert(c->out.request == c->out.request_written);
334        *sent = c->out.request;
335    }
336    pthread_mutex_unlock(&c->iolock);
337    return ret;
338}
339
340int xcb_writev(xcb_connection_t *c, struct iovec *vector, int count, uint64_t requests)
341{
342    int ret;
343    if(c->has_error)
344        return 0;
345    pthread_mutex_lock(&c->iolock);
346    c->out.request += requests;
347    ret = _xcb_out_send(c, vector, count);
348    pthread_mutex_unlock(&c->iolock);
349    return ret;
350}
351
352int xcb_flush(xcb_connection_t *c)
353{
354    int ret;
355    if(c->has_error)
356        return 0;
357    pthread_mutex_lock(&c->iolock);
358    ret = _xcb_out_flush_to(c, c->out.request);
359    pthread_mutex_unlock(&c->iolock);
360    return ret;
361}
362
363/* Private interface */
364
365int _xcb_out_init(_xcb_out *out)
366{
367    if(pthread_cond_init(&out->socket_cond, 0))
368        return 0;
369    out->return_socket = 0;
370    out->socket_closure = 0;
371    out->socket_moving = 0;
372
373    if(pthread_cond_init(&out->cond, 0))
374        return 0;
375    out->writing = 0;
376
377    out->queue_len = 0;
378
379    out->request = 0;
380    out->request_written = 0;
381
382    if(pthread_mutex_init(&out->reqlenlock, 0))
383        return 0;
384    out->maximum_request_length_tag = LAZY_NONE;
385
386    return 1;
387}
388
389void _xcb_out_destroy(_xcb_out *out)
390{
391    pthread_cond_destroy(&out->cond);
392    pthread_mutex_destroy(&out->reqlenlock);
393}
394
395int _xcb_out_send(xcb_connection_t *c, struct iovec *vector, int count)
396{
397    int ret = 1;
398    while(ret && count)
399        ret = _xcb_conn_wait(c, &c->out.cond, &vector, &count);
400    c->out.request_written = c->out.request;
401    pthread_cond_broadcast(&c->out.cond);
402    _xcb_in_wake_up_next_reader(c);
403    return ret;
404}
405
406void _xcb_out_send_sync(xcb_connection_t *c)
407{
408    prepare_socket_request(c);
409    send_sync(c);
410}
411
412int _xcb_out_flush_to(xcb_connection_t *c, uint64_t request)
413{
414    assert(XCB_SEQUENCE_COMPARE(request, <=, c->out.request));
415    if(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request))
416        return 1;
417    if(c->out.queue_len)
418    {
419        struct iovec vec;
420        vec.iov_base = c->out.queue;
421        vec.iov_len = c->out.queue_len;
422        c->out.queue_len = 0;
423        return _xcb_out_send(c, &vec, 1);
424    }
425    while(c->out.writing)
426        pthread_cond_wait(&c->out.cond, &c->iolock);
427    assert(XCB_SEQUENCE_COMPARE(c->out.request_written, >=, request));
428    return 1;
429}
430