Merge tag 'usb-3.9-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/gregkh/usb
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / net / sunrpc / auth_gss / auth_gss.c
1 /*
2 * linux/net/sunrpc/auth_gss/auth_gss.c
3 *
4 * RPCSEC_GSS client authentication.
5 *
6 * Copyright (c) 2000 The Regents of the University of Michigan.
7 * All rights reserved.
8 *
9 * Dug Song <dugsong@monkey.org>
10 * Andy Adamson <andros@umich.edu>
11 *
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
14 * are met:
15 *
16 * 1. Redistributions of source code must retain the above copyright
17 * notice, this list of conditions and the following disclaimer.
18 * 2. Redistributions in binary form must reproduce the above copyright
19 * notice, this list of conditions and the following disclaimer in the
20 * documentation and/or other materials provided with the distribution.
21 * 3. Neither the name of the University nor the names of its
22 * contributors may be used to endorse or promote products derived
23 * from this software without specific prior written permission.
24 *
25 * THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED
26 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
27 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
28 * DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
29 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
30 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
31 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
32 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
33 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
34 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
35 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
36 */
37
38
39 #include <linux/module.h>
40 #include <linux/init.h>
41 #include <linux/types.h>
42 #include <linux/slab.h>
43 #include <linux/sched.h>
44 #include <linux/pagemap.h>
45 #include <linux/sunrpc/clnt.h>
46 #include <linux/sunrpc/auth.h>
47 #include <linux/sunrpc/auth_gss.h>
48 #include <linux/sunrpc/svcauth_gss.h>
49 #include <linux/sunrpc/gss_err.h>
50 #include <linux/workqueue.h>
51 #include <linux/sunrpc/rpc_pipe_fs.h>
52 #include <linux/sunrpc/gss_api.h>
53 #include <asm/uaccess.h>
54
55 static const struct rpc_authops authgss_ops;
56
57 static const struct rpc_credops gss_credops;
58 static const struct rpc_credops gss_nullops;
59
60 #define GSS_RETRY_EXPIRED 5
61 static unsigned int gss_expired_cred_retry_delay = GSS_RETRY_EXPIRED;
62
63 #ifdef RPC_DEBUG
64 # define RPCDBG_FACILITY RPCDBG_AUTH
65 #endif
66
67 #define GSS_CRED_SLACK (RPC_MAX_AUTH_SIZE * 2)
68 /* length of a krb5 verifier (48), plus data added before arguments when
69 * using integrity (two 4-byte integers): */
70 #define GSS_VERF_SLACK 100
71
72 struct gss_auth {
73 struct kref kref;
74 struct rpc_auth rpc_auth;
75 struct gss_api_mech *mech;
76 enum rpc_gss_svc service;
77 struct rpc_clnt *client;
78 /*
79 * There are two upcall pipes; dentry[1], named "gssd", is used
80 * for the new text-based upcall; dentry[0] is named after the
81 * mechanism (for example, "krb5") and exists for
82 * backwards-compatibility with older gssd's.
83 */
84 struct rpc_pipe *pipe[2];
85 };
86
87 /* pipe_version >= 0 if and only if someone has a pipe open. */
88 static int pipe_version = -1;
89 static atomic_t pipe_users = ATOMIC_INIT(0);
90 static DEFINE_SPINLOCK(pipe_version_lock);
91 static struct rpc_wait_queue pipe_version_rpc_waitqueue;
92 static DECLARE_WAIT_QUEUE_HEAD(pipe_version_waitqueue);
93
94 static void gss_free_ctx(struct gss_cl_ctx *);
95 static const struct rpc_pipe_ops gss_upcall_ops_v0;
96 static const struct rpc_pipe_ops gss_upcall_ops_v1;
97
98 static inline struct gss_cl_ctx *
99 gss_get_ctx(struct gss_cl_ctx *ctx)
100 {
101 atomic_inc(&ctx->count);
102 return ctx;
103 }
104
105 static inline void
106 gss_put_ctx(struct gss_cl_ctx *ctx)
107 {
108 if (atomic_dec_and_test(&ctx->count))
109 gss_free_ctx(ctx);
110 }
111
112 /* gss_cred_set_ctx:
113 * called by gss_upcall_callback and gss_create_upcall in order
114 * to set the gss context. The actual exchange of an old context
115 * and a new one is protected by the pipe->lock.
116 */
117 static void
118 gss_cred_set_ctx(struct rpc_cred *cred, struct gss_cl_ctx *ctx)
119 {
120 struct gss_cred *gss_cred = container_of(cred, struct gss_cred, gc_base);
121
122 if (!test_bit(RPCAUTH_CRED_NEW, &cred->cr_flags))
123 return;
124 gss_get_ctx(ctx);
125 rcu_assign_pointer(gss_cred->gc_ctx, ctx);
126 set_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
127 smp_mb__before_clear_bit();
128 clear_bit(RPCAUTH_CRED_NEW, &cred->cr_flags);
129 }
130
131 static const void *
132 simple_get_bytes(const void *p, const void *end, void *res, size_t len)
133 {
134 const void *q = (const void *)((const char *)p + len);
135 if (unlikely(q > end || q < p))
136 return ERR_PTR(-EFAULT);
137 memcpy(res, p, len);
138 return q;
139 }
140
141 static inline const void *
142 simple_get_netobj(const void *p, const void *end, struct xdr_netobj *dest)
143 {
144 const void *q;
145 unsigned int len;
146
147 p = simple_get_bytes(p, end, &len, sizeof(len));
148 if (IS_ERR(p))
149 return p;
150 q = (const void *)((const char *)p + len);
151 if (unlikely(q > end || q < p))
152 return ERR_PTR(-EFAULT);
153 dest->data = kmemdup(p, len, GFP_NOFS);
154 if (unlikely(dest->data == NULL))
155 return ERR_PTR(-ENOMEM);
156 dest->len = len;
157 return q;
158 }
159
160 static struct gss_cl_ctx *
161 gss_cred_get_ctx(struct rpc_cred *cred)
162 {
163 struct gss_cred *gss_cred = container_of(cred, struct gss_cred, gc_base);
164 struct gss_cl_ctx *ctx = NULL;
165
166 rcu_read_lock();
167 if (gss_cred->gc_ctx)
168 ctx = gss_get_ctx(gss_cred->gc_ctx);
169 rcu_read_unlock();
170 return ctx;
171 }
172
173 static struct gss_cl_ctx *
174 gss_alloc_context(void)
175 {
176 struct gss_cl_ctx *ctx;
177
178 ctx = kzalloc(sizeof(*ctx), GFP_NOFS);
179 if (ctx != NULL) {
180 ctx->gc_proc = RPC_GSS_PROC_DATA;
181 ctx->gc_seq = 1; /* NetApp 6.4R1 doesn't accept seq. no. 0 */
182 spin_lock_init(&ctx->gc_seq_lock);
183 atomic_set(&ctx->count,1);
184 }
185 return ctx;
186 }
187
188 #define GSSD_MIN_TIMEOUT (60 * 60)
189 static const void *
190 gss_fill_context(const void *p, const void *end, struct gss_cl_ctx *ctx, struct gss_api_mech *gm)
191 {
192 const void *q;
193 unsigned int seclen;
194 unsigned int timeout;
195 unsigned long now = jiffies;
196 u32 window_size;
197 int ret;
198
199 /* First unsigned int gives the remaining lifetime in seconds of the
200 * credential - e.g. the remaining TGT lifetime for Kerberos or
201 * the -t value passed to GSSD.
202 */
203 p = simple_get_bytes(p, end, &timeout, sizeof(timeout));
204 if (IS_ERR(p))
205 goto err;
206 if (timeout == 0)
207 timeout = GSSD_MIN_TIMEOUT;
208 ctx->gc_expiry = now + ((unsigned long)timeout * HZ);
209 /* Sequence number window. Determines the maximum number of
210 * simultaneous requests
211 */
212 p = simple_get_bytes(p, end, &window_size, sizeof(window_size));
213 if (IS_ERR(p))
214 goto err;
215 ctx->gc_win = window_size;
216 /* gssd signals an error by passing ctx->gc_win = 0: */
217 if (ctx->gc_win == 0) {
218 /*
219 * in which case, p points to an error code. Anything other
220 * than -EKEYEXPIRED gets converted to -EACCES.
221 */
222 p = simple_get_bytes(p, end, &ret, sizeof(ret));
223 if (!IS_ERR(p))
224 p = (ret == -EKEYEXPIRED) ? ERR_PTR(-EKEYEXPIRED) :
225 ERR_PTR(-EACCES);
226 goto err;
227 }
228 /* copy the opaque wire context */
229 p = simple_get_netobj(p, end, &ctx->gc_wire_ctx);
230 if (IS_ERR(p))
231 goto err;
232 /* import the opaque security context */
233 p = simple_get_bytes(p, end, &seclen, sizeof(seclen));
234 if (IS_ERR(p))
235 goto err;
236 q = (const void *)((const char *)p + seclen);
237 if (unlikely(q > end || q < p)) {
238 p = ERR_PTR(-EFAULT);
239 goto err;
240 }
241 ret = gss_import_sec_context(p, seclen, gm, &ctx->gc_gss_ctx, GFP_NOFS);
242 if (ret < 0) {
243 p = ERR_PTR(ret);
244 goto err;
245 }
246 dprintk("RPC: %s Success. gc_expiry %lu now %lu timeout %u\n",
247 __func__, ctx->gc_expiry, now, timeout);
248 return q;
249 err:
250 dprintk("RPC: %s returns error %ld\n", __func__, -PTR_ERR(p));
251 return p;
252 }
253
254 #define UPCALL_BUF_LEN 128
255
256 struct gss_upcall_msg {
257 atomic_t count;
258 uid_t uid;
259 struct rpc_pipe_msg msg;
260 struct list_head list;
261 struct gss_auth *auth;
262 struct rpc_pipe *pipe;
263 struct rpc_wait_queue rpc_waitqueue;
264 wait_queue_head_t waitqueue;
265 struct gss_cl_ctx *ctx;
266 char databuf[UPCALL_BUF_LEN];
267 };
268
269 static int get_pipe_version(void)
270 {
271 int ret;
272
273 spin_lock(&pipe_version_lock);
274 if (pipe_version >= 0) {
275 atomic_inc(&pipe_users);
276 ret = pipe_version;
277 } else
278 ret = -EAGAIN;
279 spin_unlock(&pipe_version_lock);
280 return ret;
281 }
282
283 static void put_pipe_version(void)
284 {
285 if (atomic_dec_and_lock(&pipe_users, &pipe_version_lock)) {
286 pipe_version = -1;
287 spin_unlock(&pipe_version_lock);
288 }
289 }
290
291 static void
292 gss_release_msg(struct gss_upcall_msg *gss_msg)
293 {
294 if (!atomic_dec_and_test(&gss_msg->count))
295 return;
296 put_pipe_version();
297 BUG_ON(!list_empty(&gss_msg->list));
298 if (gss_msg->ctx != NULL)
299 gss_put_ctx(gss_msg->ctx);
300 rpc_destroy_wait_queue(&gss_msg->rpc_waitqueue);
301 kfree(gss_msg);
302 }
303
304 static struct gss_upcall_msg *
305 __gss_find_upcall(struct rpc_pipe *pipe, uid_t uid)
306 {
307 struct gss_upcall_msg *pos;
308 list_for_each_entry(pos, &pipe->in_downcall, list) {
309 if (pos->uid != uid)
310 continue;
311 atomic_inc(&pos->count);
312 dprintk("RPC: %s found msg %p\n", __func__, pos);
313 return pos;
314 }
315 dprintk("RPC: %s found nothing\n", __func__);
316 return NULL;
317 }
318
319 /* Try to add an upcall to the pipefs queue.
320 * If an upcall owned by our uid already exists, then we return a reference
321 * to that upcall instead of adding the new upcall.
322 */
323 static inline struct gss_upcall_msg *
324 gss_add_msg(struct gss_upcall_msg *gss_msg)
325 {
326 struct rpc_pipe *pipe = gss_msg->pipe;
327 struct gss_upcall_msg *old;
328
329 spin_lock(&pipe->lock);
330 old = __gss_find_upcall(pipe, gss_msg->uid);
331 if (old == NULL) {
332 atomic_inc(&gss_msg->count);
333 list_add(&gss_msg->list, &pipe->in_downcall);
334 } else
335 gss_msg = old;
336 spin_unlock(&pipe->lock);
337 return gss_msg;
338 }
339
340 static void
341 __gss_unhash_msg(struct gss_upcall_msg *gss_msg)
342 {
343 list_del_init(&gss_msg->list);
344 rpc_wake_up_status(&gss_msg->rpc_waitqueue, gss_msg->msg.errno);
345 wake_up_all(&gss_msg->waitqueue);
346 atomic_dec(&gss_msg->count);
347 }
348
349 static void
350 gss_unhash_msg(struct gss_upcall_msg *gss_msg)
351 {
352 struct rpc_pipe *pipe = gss_msg->pipe;
353
354 if (list_empty(&gss_msg->list))
355 return;
356 spin_lock(&pipe->lock);
357 if (!list_empty(&gss_msg->list))
358 __gss_unhash_msg(gss_msg);
359 spin_unlock(&pipe->lock);
360 }
361
362 static void
363 gss_handle_downcall_result(struct gss_cred *gss_cred, struct gss_upcall_msg *gss_msg)
364 {
365 switch (gss_msg->msg.errno) {
366 case 0:
367 if (gss_msg->ctx == NULL)
368 break;
369 clear_bit(RPCAUTH_CRED_NEGATIVE, &gss_cred->gc_base.cr_flags);
370 gss_cred_set_ctx(&gss_cred->gc_base, gss_msg->ctx);
371 break;
372 case -EKEYEXPIRED:
373 set_bit(RPCAUTH_CRED_NEGATIVE, &gss_cred->gc_base.cr_flags);
374 }
375 gss_cred->gc_upcall_timestamp = jiffies;
376 gss_cred->gc_upcall = NULL;
377 rpc_wake_up_status(&gss_msg->rpc_waitqueue, gss_msg->msg.errno);
378 }
379
380 static void
381 gss_upcall_callback(struct rpc_task *task)
382 {
383 struct gss_cred *gss_cred = container_of(task->tk_rqstp->rq_cred,
384 struct gss_cred, gc_base);
385 struct gss_upcall_msg *gss_msg = gss_cred->gc_upcall;
386 struct rpc_pipe *pipe = gss_msg->pipe;
387
388 spin_lock(&pipe->lock);
389 gss_handle_downcall_result(gss_cred, gss_msg);
390 spin_unlock(&pipe->lock);
391 task->tk_status = gss_msg->msg.errno;
392 gss_release_msg(gss_msg);
393 }
394
395 static void gss_encode_v0_msg(struct gss_upcall_msg *gss_msg)
396 {
397 gss_msg->msg.data = &gss_msg->uid;
398 gss_msg->msg.len = sizeof(gss_msg->uid);
399 }
400
401 static void gss_encode_v1_msg(struct gss_upcall_msg *gss_msg,
402 struct rpc_clnt *clnt,
403 const char *service_name)
404 {
405 struct gss_api_mech *mech = gss_msg->auth->mech;
406 char *p = gss_msg->databuf;
407 int len = 0;
408
409 gss_msg->msg.len = sprintf(gss_msg->databuf, "mech=%s uid=%d ",
410 mech->gm_name,
411 gss_msg->uid);
412 p += gss_msg->msg.len;
413 if (clnt->cl_principal) {
414 len = sprintf(p, "target=%s ", clnt->cl_principal);
415 p += len;
416 gss_msg->msg.len += len;
417 }
418 if (service_name != NULL) {
419 len = sprintf(p, "service=%s ", service_name);
420 p += len;
421 gss_msg->msg.len += len;
422 }
423 if (mech->gm_upcall_enctypes) {
424 len = sprintf(p, "enctypes=%s ", mech->gm_upcall_enctypes);
425 p += len;
426 gss_msg->msg.len += len;
427 }
428 len = sprintf(p, "\n");
429 gss_msg->msg.len += len;
430
431 gss_msg->msg.data = gss_msg->databuf;
432 BUG_ON(gss_msg->msg.len > UPCALL_BUF_LEN);
433 }
434
435 static void gss_encode_msg(struct gss_upcall_msg *gss_msg,
436 struct rpc_clnt *clnt,
437 const char *service_name)
438 {
439 if (pipe_version == 0)
440 gss_encode_v0_msg(gss_msg);
441 else /* pipe_version == 1 */
442 gss_encode_v1_msg(gss_msg, clnt, service_name);
443 }
444
445 static struct gss_upcall_msg *
446 gss_alloc_msg(struct gss_auth *gss_auth, struct rpc_clnt *clnt,
447 uid_t uid, const char *service_name)
448 {
449 struct gss_upcall_msg *gss_msg;
450 int vers;
451
452 gss_msg = kzalloc(sizeof(*gss_msg), GFP_NOFS);
453 if (gss_msg == NULL)
454 return ERR_PTR(-ENOMEM);
455 vers = get_pipe_version();
456 if (vers < 0) {
457 kfree(gss_msg);
458 return ERR_PTR(vers);
459 }
460 gss_msg->pipe = gss_auth->pipe[vers];
461 INIT_LIST_HEAD(&gss_msg->list);
462 rpc_init_wait_queue(&gss_msg->rpc_waitqueue, "RPCSEC_GSS upcall waitq");
463 init_waitqueue_head(&gss_msg->waitqueue);
464 atomic_set(&gss_msg->count, 1);
465 gss_msg->uid = uid;
466 gss_msg->auth = gss_auth;
467 gss_encode_msg(gss_msg, clnt, service_name);
468 return gss_msg;
469 }
470
471 static struct gss_upcall_msg *
472 gss_setup_upcall(struct rpc_clnt *clnt, struct gss_auth *gss_auth, struct rpc_cred *cred)
473 {
474 struct gss_cred *gss_cred = container_of(cred,
475 struct gss_cred, gc_base);
476 struct gss_upcall_msg *gss_new, *gss_msg;
477 uid_t uid = cred->cr_uid;
478
479 gss_new = gss_alloc_msg(gss_auth, clnt, uid, gss_cred->gc_principal);
480 if (IS_ERR(gss_new))
481 return gss_new;
482 gss_msg = gss_add_msg(gss_new);
483 if (gss_msg == gss_new) {
484 int res = rpc_queue_upcall(gss_new->pipe, &gss_new->msg);
485 if (res) {
486 gss_unhash_msg(gss_new);
487 gss_msg = ERR_PTR(res);
488 }
489 } else
490 gss_release_msg(gss_new);
491 return gss_msg;
492 }
493
494 static void warn_gssd(void)
495 {
496 static unsigned long ratelimit;
497 unsigned long now = jiffies;
498
499 if (time_after(now, ratelimit)) {
500 printk(KERN_WARNING "RPC: AUTH_GSS upcall timed out.\n"
501 "Please check user daemon is running.\n");
502 ratelimit = now + 15*HZ;
503 }
504 }
505
506 static inline int
507 gss_refresh_upcall(struct rpc_task *task)
508 {
509 struct rpc_cred *cred = task->tk_rqstp->rq_cred;
510 struct gss_auth *gss_auth = container_of(cred->cr_auth,
511 struct gss_auth, rpc_auth);
512 struct gss_cred *gss_cred = container_of(cred,
513 struct gss_cred, gc_base);
514 struct gss_upcall_msg *gss_msg;
515 struct rpc_pipe *pipe;
516 int err = 0;
517
518 dprintk("RPC: %5u %s for uid %u\n",
519 task->tk_pid, __func__, cred->cr_uid);
520 gss_msg = gss_setup_upcall(task->tk_client, gss_auth, cred);
521 if (PTR_ERR(gss_msg) == -EAGAIN) {
522 /* XXX: warning on the first, under the assumption we
523 * shouldn't normally hit this case on a refresh. */
524 warn_gssd();
525 task->tk_timeout = 15*HZ;
526 rpc_sleep_on(&pipe_version_rpc_waitqueue, task, NULL);
527 return -EAGAIN;
528 }
529 if (IS_ERR(gss_msg)) {
530 err = PTR_ERR(gss_msg);
531 goto out;
532 }
533 pipe = gss_msg->pipe;
534 spin_lock(&pipe->lock);
535 if (gss_cred->gc_upcall != NULL)
536 rpc_sleep_on(&gss_cred->gc_upcall->rpc_waitqueue, task, NULL);
537 else if (gss_msg->ctx == NULL && gss_msg->msg.errno >= 0) {
538 task->tk_timeout = 0;
539 gss_cred->gc_upcall = gss_msg;
540 /* gss_upcall_callback will release the reference to gss_upcall_msg */
541 atomic_inc(&gss_msg->count);
542 rpc_sleep_on(&gss_msg->rpc_waitqueue, task, gss_upcall_callback);
543 } else {
544 gss_handle_downcall_result(gss_cred, gss_msg);
545 err = gss_msg->msg.errno;
546 }
547 spin_unlock(&pipe->lock);
548 gss_release_msg(gss_msg);
549 out:
550 dprintk("RPC: %5u %s for uid %u result %d\n",
551 task->tk_pid, __func__, cred->cr_uid, err);
552 return err;
553 }
554
555 static inline int
556 gss_create_upcall(struct gss_auth *gss_auth, struct gss_cred *gss_cred)
557 {
558 struct rpc_pipe *pipe;
559 struct rpc_cred *cred = &gss_cred->gc_base;
560 struct gss_upcall_msg *gss_msg;
561 DEFINE_WAIT(wait);
562 int err = 0;
563
564 dprintk("RPC: %s for uid %u\n", __func__, cred->cr_uid);
565 retry:
566 gss_msg = gss_setup_upcall(gss_auth->client, gss_auth, cred);
567 if (PTR_ERR(gss_msg) == -EAGAIN) {
568 err = wait_event_interruptible_timeout(pipe_version_waitqueue,
569 pipe_version >= 0, 15*HZ);
570 if (pipe_version < 0) {
571 warn_gssd();
572 err = -EACCES;
573 }
574 if (err)
575 goto out;
576 goto retry;
577 }
578 if (IS_ERR(gss_msg)) {
579 err = PTR_ERR(gss_msg);
580 goto out;
581 }
582 pipe = gss_msg->pipe;
583 for (;;) {
584 prepare_to_wait(&gss_msg->waitqueue, &wait, TASK_KILLABLE);
585 spin_lock(&pipe->lock);
586 if (gss_msg->ctx != NULL || gss_msg->msg.errno < 0) {
587 break;
588 }
589 spin_unlock(&pipe->lock);
590 if (fatal_signal_pending(current)) {
591 err = -ERESTARTSYS;
592 goto out_intr;
593 }
594 schedule();
595 }
596 if (gss_msg->ctx)
597 gss_cred_set_ctx(cred, gss_msg->ctx);
598 else
599 err = gss_msg->msg.errno;
600 spin_unlock(&pipe->lock);
601 out_intr:
602 finish_wait(&gss_msg->waitqueue, &wait);
603 gss_release_msg(gss_msg);
604 out:
605 dprintk("RPC: %s for uid %u result %d\n",
606 __func__, cred->cr_uid, err);
607 return err;
608 }
609
610 #define MSG_BUF_MAXSIZE 1024
611
612 static ssize_t
613 gss_pipe_downcall(struct file *filp, const char __user *src, size_t mlen)
614 {
615 const void *p, *end;
616 void *buf;
617 struct gss_upcall_msg *gss_msg;
618 struct rpc_pipe *pipe = RPC_I(filp->f_dentry->d_inode)->pipe;
619 struct gss_cl_ctx *ctx;
620 uid_t uid;
621 ssize_t err = -EFBIG;
622
623 if (mlen > MSG_BUF_MAXSIZE)
624 goto out;
625 err = -ENOMEM;
626 buf = kmalloc(mlen, GFP_NOFS);
627 if (!buf)
628 goto out;
629
630 err = -EFAULT;
631 if (copy_from_user(buf, src, mlen))
632 goto err;
633
634 end = (const void *)((char *)buf + mlen);
635 p = simple_get_bytes(buf, end, &uid, sizeof(uid));
636 if (IS_ERR(p)) {
637 err = PTR_ERR(p);
638 goto err;
639 }
640
641 err = -ENOMEM;
642 ctx = gss_alloc_context();
643 if (ctx == NULL)
644 goto err;
645
646 err = -ENOENT;
647 /* Find a matching upcall */
648 spin_lock(&pipe->lock);
649 gss_msg = __gss_find_upcall(pipe, uid);
650 if (gss_msg == NULL) {
651 spin_unlock(&pipe->lock);
652 goto err_put_ctx;
653 }
654 list_del_init(&gss_msg->list);
655 spin_unlock(&pipe->lock);
656
657 p = gss_fill_context(p, end, ctx, gss_msg->auth->mech);
658 if (IS_ERR(p)) {
659 err = PTR_ERR(p);
660 switch (err) {
661 case -EACCES:
662 case -EKEYEXPIRED:
663 gss_msg->msg.errno = err;
664 err = mlen;
665 break;
666 case -EFAULT:
667 case -ENOMEM:
668 case -EINVAL:
669 case -ENOSYS:
670 gss_msg->msg.errno = -EAGAIN;
671 break;
672 default:
673 printk(KERN_CRIT "%s: bad return from "
674 "gss_fill_context: %zd\n", __func__, err);
675 BUG();
676 }
677 goto err_release_msg;
678 }
679 gss_msg->ctx = gss_get_ctx(ctx);
680 err = mlen;
681
682 err_release_msg:
683 spin_lock(&pipe->lock);
684 __gss_unhash_msg(gss_msg);
685 spin_unlock(&pipe->lock);
686 gss_release_msg(gss_msg);
687 err_put_ctx:
688 gss_put_ctx(ctx);
689 err:
690 kfree(buf);
691 out:
692 dprintk("RPC: %s returning %Zd\n", __func__, err);
693 return err;
694 }
695
696 static int gss_pipe_open(struct inode *inode, int new_version)
697 {
698 int ret = 0;
699
700 spin_lock(&pipe_version_lock);
701 if (pipe_version < 0) {
702 /* First open of any gss pipe determines the version: */
703 pipe_version = new_version;
704 rpc_wake_up(&pipe_version_rpc_waitqueue);
705 wake_up(&pipe_version_waitqueue);
706 } else if (pipe_version != new_version) {
707 /* Trying to open a pipe of a different version */
708 ret = -EBUSY;
709 goto out;
710 }
711 atomic_inc(&pipe_users);
712 out:
713 spin_unlock(&pipe_version_lock);
714 return ret;
715
716 }
717
718 static int gss_pipe_open_v0(struct inode *inode)
719 {
720 return gss_pipe_open(inode, 0);
721 }
722
723 static int gss_pipe_open_v1(struct inode *inode)
724 {
725 return gss_pipe_open(inode, 1);
726 }
727
728 static void
729 gss_pipe_release(struct inode *inode)
730 {
731 struct rpc_pipe *pipe = RPC_I(inode)->pipe;
732 struct gss_upcall_msg *gss_msg;
733
734 restart:
735 spin_lock(&pipe->lock);
736 list_for_each_entry(gss_msg, &pipe->in_downcall, list) {
737
738 if (!list_empty(&gss_msg->msg.list))
739 continue;
740 gss_msg->msg.errno = -EPIPE;
741 atomic_inc(&gss_msg->count);
742 __gss_unhash_msg(gss_msg);
743 spin_unlock(&pipe->lock);
744 gss_release_msg(gss_msg);
745 goto restart;
746 }
747 spin_unlock(&pipe->lock);
748
749 put_pipe_version();
750 }
751
752 static void
753 gss_pipe_destroy_msg(struct rpc_pipe_msg *msg)
754 {
755 struct gss_upcall_msg *gss_msg = container_of(msg, struct gss_upcall_msg, msg);
756
757 if (msg->errno < 0) {
758 dprintk("RPC: %s releasing msg %p\n",
759 __func__, gss_msg);
760 atomic_inc(&gss_msg->count);
761 gss_unhash_msg(gss_msg);
762 if (msg->errno == -ETIMEDOUT)
763 warn_gssd();
764 gss_release_msg(gss_msg);
765 }
766 }
767
768 static void gss_pipes_dentries_destroy(struct rpc_auth *auth)
769 {
770 struct gss_auth *gss_auth;
771
772 gss_auth = container_of(auth, struct gss_auth, rpc_auth);
773 if (gss_auth->pipe[0]->dentry)
774 rpc_unlink(gss_auth->pipe[0]->dentry);
775 if (gss_auth->pipe[1]->dentry)
776 rpc_unlink(gss_auth->pipe[1]->dentry);
777 }
778
779 static int gss_pipes_dentries_create(struct rpc_auth *auth)
780 {
781 int err;
782 struct gss_auth *gss_auth;
783 struct rpc_clnt *clnt;
784
785 gss_auth = container_of(auth, struct gss_auth, rpc_auth);
786 clnt = gss_auth->client;
787
788 gss_auth->pipe[1]->dentry = rpc_mkpipe_dentry(clnt->cl_dentry,
789 "gssd",
790 clnt, gss_auth->pipe[1]);
791 if (IS_ERR(gss_auth->pipe[1]->dentry))
792 return PTR_ERR(gss_auth->pipe[1]->dentry);
793 gss_auth->pipe[0]->dentry = rpc_mkpipe_dentry(clnt->cl_dentry,
794 gss_auth->mech->gm_name,
795 clnt, gss_auth->pipe[0]);
796 if (IS_ERR(gss_auth->pipe[0]->dentry)) {
797 err = PTR_ERR(gss_auth->pipe[0]->dentry);
798 goto err_unlink_pipe_1;
799 }
800 return 0;
801
802 err_unlink_pipe_1:
803 rpc_unlink(gss_auth->pipe[1]->dentry);
804 return err;
805 }
806
807 static void gss_pipes_dentries_destroy_net(struct rpc_clnt *clnt,
808 struct rpc_auth *auth)
809 {
810 struct net *net = rpc_net_ns(clnt);
811 struct super_block *sb;
812
813 sb = rpc_get_sb_net(net);
814 if (sb) {
815 if (clnt->cl_dentry)
816 gss_pipes_dentries_destroy(auth);
817 rpc_put_sb_net(net);
818 }
819 }
820
821 static int gss_pipes_dentries_create_net(struct rpc_clnt *clnt,
822 struct rpc_auth *auth)
823 {
824 struct net *net = rpc_net_ns(clnt);
825 struct super_block *sb;
826 int err = 0;
827
828 sb = rpc_get_sb_net(net);
829 if (sb) {
830 if (clnt->cl_dentry)
831 err = gss_pipes_dentries_create(auth);
832 rpc_put_sb_net(net);
833 }
834 return err;
835 }
836
837 /*
838 * NOTE: we have the opportunity to use different
839 * parameters based on the input flavor (which must be a pseudoflavor)
840 */
841 static struct rpc_auth *
842 gss_create(struct rpc_clnt *clnt, rpc_authflavor_t flavor)
843 {
844 struct gss_auth *gss_auth;
845 struct rpc_auth * auth;
846 int err = -ENOMEM; /* XXX? */
847
848 dprintk("RPC: creating GSS authenticator for client %p\n", clnt);
849
850 if (!try_module_get(THIS_MODULE))
851 return ERR_PTR(err);
852 if (!(gss_auth = kmalloc(sizeof(*gss_auth), GFP_KERNEL)))
853 goto out_dec;
854 gss_auth->client = clnt;
855 err = -EINVAL;
856 gss_auth->mech = gss_mech_get_by_pseudoflavor(flavor);
857 if (!gss_auth->mech) {
858 printk(KERN_WARNING "%s: Pseudoflavor %d not found!\n",
859 __func__, flavor);
860 goto err_free;
861 }
862 gss_auth->service = gss_pseudoflavor_to_service(gss_auth->mech, flavor);
863 if (gss_auth->service == 0)
864 goto err_put_mech;
865 auth = &gss_auth->rpc_auth;
866 auth->au_cslack = GSS_CRED_SLACK >> 2;
867 auth->au_rslack = GSS_VERF_SLACK >> 2;
868 auth->au_ops = &authgss_ops;
869 auth->au_flavor = flavor;
870 atomic_set(&auth->au_count, 1);
871 kref_init(&gss_auth->kref);
872
873 /*
874 * Note: if we created the old pipe first, then someone who
875 * examined the directory at the right moment might conclude
876 * that we supported only the old pipe. So we instead create
877 * the new pipe first.
878 */
879 gss_auth->pipe[1] = rpc_mkpipe_data(&gss_upcall_ops_v1,
880 RPC_PIPE_WAIT_FOR_OPEN);
881 if (IS_ERR(gss_auth->pipe[1])) {
882 err = PTR_ERR(gss_auth->pipe[1]);
883 goto err_put_mech;
884 }
885
886 gss_auth->pipe[0] = rpc_mkpipe_data(&gss_upcall_ops_v0,
887 RPC_PIPE_WAIT_FOR_OPEN);
888 if (IS_ERR(gss_auth->pipe[0])) {
889 err = PTR_ERR(gss_auth->pipe[0]);
890 goto err_destroy_pipe_1;
891 }
892 err = gss_pipes_dentries_create_net(clnt, auth);
893 if (err)
894 goto err_destroy_pipe_0;
895 err = rpcauth_init_credcache(auth);
896 if (err)
897 goto err_unlink_pipes;
898
899 return auth;
900 err_unlink_pipes:
901 gss_pipes_dentries_destroy_net(clnt, auth);
902 err_destroy_pipe_0:
903 rpc_destroy_pipe_data(gss_auth->pipe[0]);
904 err_destroy_pipe_1:
905 rpc_destroy_pipe_data(gss_auth->pipe[1]);
906 err_put_mech:
907 gss_mech_put(gss_auth->mech);
908 err_free:
909 kfree(gss_auth);
910 out_dec:
911 module_put(THIS_MODULE);
912 return ERR_PTR(err);
913 }
914
915 static void
916 gss_free(struct gss_auth *gss_auth)
917 {
918 gss_pipes_dentries_destroy_net(gss_auth->client, &gss_auth->rpc_auth);
919 rpc_destroy_pipe_data(gss_auth->pipe[0]);
920 rpc_destroy_pipe_data(gss_auth->pipe[1]);
921 gss_mech_put(gss_auth->mech);
922
923 kfree(gss_auth);
924 module_put(THIS_MODULE);
925 }
926
927 static void
928 gss_free_callback(struct kref *kref)
929 {
930 struct gss_auth *gss_auth = container_of(kref, struct gss_auth, kref);
931
932 gss_free(gss_auth);
933 }
934
935 static void
936 gss_destroy(struct rpc_auth *auth)
937 {
938 struct gss_auth *gss_auth;
939
940 dprintk("RPC: destroying GSS authenticator %p flavor %d\n",
941 auth, auth->au_flavor);
942
943 rpcauth_destroy_credcache(auth);
944
945 gss_auth = container_of(auth, struct gss_auth, rpc_auth);
946 kref_put(&gss_auth->kref, gss_free_callback);
947 }
948
949 /*
950 * gss_destroying_context will cause the RPCSEC_GSS to send a NULL RPC call
951 * to the server with the GSS control procedure field set to
952 * RPC_GSS_PROC_DESTROY. This should normally cause the server to release
953 * all RPCSEC_GSS state associated with that context.
954 */
955 static int
956 gss_destroying_context(struct rpc_cred *cred)
957 {
958 struct gss_cred *gss_cred = container_of(cred, struct gss_cred, gc_base);
959 struct gss_auth *gss_auth = container_of(cred->cr_auth, struct gss_auth, rpc_auth);
960 struct rpc_task *task;
961
962 if (gss_cred->gc_ctx == NULL ||
963 test_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags) == 0)
964 return 0;
965
966 gss_cred->gc_ctx->gc_proc = RPC_GSS_PROC_DESTROY;
967 cred->cr_ops = &gss_nullops;
968
969 /* Take a reference to ensure the cred will be destroyed either
970 * by the RPC call or by the put_rpccred() below */
971 get_rpccred(cred);
972
973 task = rpc_call_null(gss_auth->client, cred, RPC_TASK_ASYNC|RPC_TASK_SOFT);
974 if (!IS_ERR(task))
975 rpc_put_task(task);
976
977 put_rpccred(cred);
978 return 1;
979 }
980
981 /* gss_destroy_cred (and gss_free_ctx) are used to clean up after failure
982 * to create a new cred or context, so they check that things have been
983 * allocated before freeing them. */
984 static void
985 gss_do_free_ctx(struct gss_cl_ctx *ctx)
986 {
987 dprintk("RPC: %s\n", __func__);
988
989 gss_delete_sec_context(&ctx->gc_gss_ctx);
990 kfree(ctx->gc_wire_ctx.data);
991 kfree(ctx);
992 }
993
994 static void
995 gss_free_ctx_callback(struct rcu_head *head)
996 {
997 struct gss_cl_ctx *ctx = container_of(head, struct gss_cl_ctx, gc_rcu);
998 gss_do_free_ctx(ctx);
999 }
1000
1001 static void
1002 gss_free_ctx(struct gss_cl_ctx *ctx)
1003 {
1004 call_rcu(&ctx->gc_rcu, gss_free_ctx_callback);
1005 }
1006
1007 static void
1008 gss_free_cred(struct gss_cred *gss_cred)
1009 {
1010 dprintk("RPC: %s cred=%p\n", __func__, gss_cred);
1011 kfree(gss_cred);
1012 }
1013
1014 static void
1015 gss_free_cred_callback(struct rcu_head *head)
1016 {
1017 struct gss_cred *gss_cred = container_of(head, struct gss_cred, gc_base.cr_rcu);
1018 gss_free_cred(gss_cred);
1019 }
1020
1021 static void
1022 gss_destroy_nullcred(struct rpc_cred *cred)
1023 {
1024 struct gss_cred *gss_cred = container_of(cred, struct gss_cred, gc_base);
1025 struct gss_auth *gss_auth = container_of(cred->cr_auth, struct gss_auth, rpc_auth);
1026 struct gss_cl_ctx *ctx = gss_cred->gc_ctx;
1027
1028 RCU_INIT_POINTER(gss_cred->gc_ctx, NULL);
1029 call_rcu(&cred->cr_rcu, gss_free_cred_callback);
1030 if (ctx)
1031 gss_put_ctx(ctx);
1032 kref_put(&gss_auth->kref, gss_free_callback);
1033 }
1034
1035 static void
1036 gss_destroy_cred(struct rpc_cred *cred)
1037 {
1038
1039 if (gss_destroying_context(cred))
1040 return;
1041 gss_destroy_nullcred(cred);
1042 }
1043
1044 /*
1045 * Lookup RPCSEC_GSS cred for the current process
1046 */
1047 static struct rpc_cred *
1048 gss_lookup_cred(struct rpc_auth *auth, struct auth_cred *acred, int flags)
1049 {
1050 return rpcauth_lookup_credcache(auth, acred, flags);
1051 }
1052
1053 static struct rpc_cred *
1054 gss_create_cred(struct rpc_auth *auth, struct auth_cred *acred, int flags)
1055 {
1056 struct gss_auth *gss_auth = container_of(auth, struct gss_auth, rpc_auth);
1057 struct gss_cred *cred = NULL;
1058 int err = -ENOMEM;
1059
1060 dprintk("RPC: %s for uid %d, flavor %d\n",
1061 __func__, acred->uid, auth->au_flavor);
1062
1063 if (!(cred = kzalloc(sizeof(*cred), GFP_NOFS)))
1064 goto out_err;
1065
1066 rpcauth_init_cred(&cred->gc_base, acred, auth, &gss_credops);
1067 /*
1068 * Note: in order to force a call to call_refresh(), we deliberately
1069 * fail to flag the credential as RPCAUTH_CRED_UPTODATE.
1070 */
1071 cred->gc_base.cr_flags = 1UL << RPCAUTH_CRED_NEW;
1072 cred->gc_service = gss_auth->service;
1073 cred->gc_principal = NULL;
1074 if (acred->machine_cred)
1075 cred->gc_principal = acred->principal;
1076 kref_get(&gss_auth->kref);
1077 return &cred->gc_base;
1078
1079 out_err:
1080 dprintk("RPC: %s failed with error %d\n", __func__, err);
1081 return ERR_PTR(err);
1082 }
1083
1084 static int
1085 gss_cred_init(struct rpc_auth *auth, struct rpc_cred *cred)
1086 {
1087 struct gss_auth *gss_auth = container_of(auth, struct gss_auth, rpc_auth);
1088 struct gss_cred *gss_cred = container_of(cred,struct gss_cred, gc_base);
1089 int err;
1090
1091 do {
1092 err = gss_create_upcall(gss_auth, gss_cred);
1093 } while (err == -EAGAIN);
1094 return err;
1095 }
1096
1097 static int
1098 gss_match(struct auth_cred *acred, struct rpc_cred *rc, int flags)
1099 {
1100 struct gss_cred *gss_cred = container_of(rc, struct gss_cred, gc_base);
1101
1102 if (test_bit(RPCAUTH_CRED_NEW, &rc->cr_flags))
1103 goto out;
1104 /* Don't match with creds that have expired. */
1105 if (time_after(jiffies, gss_cred->gc_ctx->gc_expiry))
1106 return 0;
1107 if (!test_bit(RPCAUTH_CRED_UPTODATE, &rc->cr_flags))
1108 return 0;
1109 out:
1110 if (acred->principal != NULL) {
1111 if (gss_cred->gc_principal == NULL)
1112 return 0;
1113 return strcmp(acred->principal, gss_cred->gc_principal) == 0;
1114 }
1115 if (gss_cred->gc_principal != NULL)
1116 return 0;
1117 return rc->cr_uid == acred->uid;
1118 }
1119
1120 /*
1121 * Marshal credentials.
1122 * Maybe we should keep a cached credential for performance reasons.
1123 */
1124 static __be32 *
1125 gss_marshal(struct rpc_task *task, __be32 *p)
1126 {
1127 struct rpc_rqst *req = task->tk_rqstp;
1128 struct rpc_cred *cred = req->rq_cred;
1129 struct gss_cred *gss_cred = container_of(cred, struct gss_cred,
1130 gc_base);
1131 struct gss_cl_ctx *ctx = gss_cred_get_ctx(cred);
1132 __be32 *cred_len;
1133 u32 maj_stat = 0;
1134 struct xdr_netobj mic;
1135 struct kvec iov;
1136 struct xdr_buf verf_buf;
1137
1138 dprintk("RPC: %5u %s\n", task->tk_pid, __func__);
1139
1140 *p++ = htonl(RPC_AUTH_GSS);
1141 cred_len = p++;
1142
1143 spin_lock(&ctx->gc_seq_lock);
1144 req->rq_seqno = ctx->gc_seq++;
1145 spin_unlock(&ctx->gc_seq_lock);
1146
1147 *p++ = htonl((u32) RPC_GSS_VERSION);
1148 *p++ = htonl((u32) ctx->gc_proc);
1149 *p++ = htonl((u32) req->rq_seqno);
1150 *p++ = htonl((u32) gss_cred->gc_service);
1151 p = xdr_encode_netobj(p, &ctx->gc_wire_ctx);
1152 *cred_len = htonl((p - (cred_len + 1)) << 2);
1153
1154 /* We compute the checksum for the verifier over the xdr-encoded bytes
1155 * starting with the xid and ending at the end of the credential: */
1156 iov.iov_base = xprt_skip_transport_header(req->rq_xprt,
1157 req->rq_snd_buf.head[0].iov_base);
1158 iov.iov_len = (u8 *)p - (u8 *)iov.iov_base;
1159 xdr_buf_from_iov(&iov, &verf_buf);
1160
1161 /* set verifier flavor*/
1162 *p++ = htonl(RPC_AUTH_GSS);
1163
1164 mic.data = (u8 *)(p + 1);
1165 maj_stat = gss_get_mic(ctx->gc_gss_ctx, &verf_buf, &mic);
1166 if (maj_stat == GSS_S_CONTEXT_EXPIRED) {
1167 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1168 } else if (maj_stat != 0) {
1169 printk("gss_marshal: gss_get_mic FAILED (%d)\n", maj_stat);
1170 goto out_put_ctx;
1171 }
1172 p = xdr_encode_opaque(p, NULL, mic.len);
1173 gss_put_ctx(ctx);
1174 return p;
1175 out_put_ctx:
1176 gss_put_ctx(ctx);
1177 return NULL;
1178 }
1179
1180 static int gss_renew_cred(struct rpc_task *task)
1181 {
1182 struct rpc_cred *oldcred = task->tk_rqstp->rq_cred;
1183 struct gss_cred *gss_cred = container_of(oldcred,
1184 struct gss_cred,
1185 gc_base);
1186 struct rpc_auth *auth = oldcred->cr_auth;
1187 struct auth_cred acred = {
1188 .uid = oldcred->cr_uid,
1189 .principal = gss_cred->gc_principal,
1190 .machine_cred = (gss_cred->gc_principal != NULL ? 1 : 0),
1191 };
1192 struct rpc_cred *new;
1193
1194 new = gss_lookup_cred(auth, &acred, RPCAUTH_LOOKUP_NEW);
1195 if (IS_ERR(new))
1196 return PTR_ERR(new);
1197 task->tk_rqstp->rq_cred = new;
1198 put_rpccred(oldcred);
1199 return 0;
1200 }
1201
1202 static int gss_cred_is_negative_entry(struct rpc_cred *cred)
1203 {
1204 if (test_bit(RPCAUTH_CRED_NEGATIVE, &cred->cr_flags)) {
1205 unsigned long now = jiffies;
1206 unsigned long begin, expire;
1207 struct gss_cred *gss_cred;
1208
1209 gss_cred = container_of(cred, struct gss_cred, gc_base);
1210 begin = gss_cred->gc_upcall_timestamp;
1211 expire = begin + gss_expired_cred_retry_delay * HZ;
1212
1213 if (time_in_range_open(now, begin, expire))
1214 return 1;
1215 }
1216 return 0;
1217 }
1218
1219 /*
1220 * Refresh credentials. XXX - finish
1221 */
1222 static int
1223 gss_refresh(struct rpc_task *task)
1224 {
1225 struct rpc_cred *cred = task->tk_rqstp->rq_cred;
1226 int ret = 0;
1227
1228 if (gss_cred_is_negative_entry(cred))
1229 return -EKEYEXPIRED;
1230
1231 if (!test_bit(RPCAUTH_CRED_NEW, &cred->cr_flags) &&
1232 !test_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags)) {
1233 ret = gss_renew_cred(task);
1234 if (ret < 0)
1235 goto out;
1236 cred = task->tk_rqstp->rq_cred;
1237 }
1238
1239 if (test_bit(RPCAUTH_CRED_NEW, &cred->cr_flags))
1240 ret = gss_refresh_upcall(task);
1241 out:
1242 return ret;
1243 }
1244
1245 /* Dummy refresh routine: used only when destroying the context */
1246 static int
1247 gss_refresh_null(struct rpc_task *task)
1248 {
1249 return -EACCES;
1250 }
1251
1252 static __be32 *
1253 gss_validate(struct rpc_task *task, __be32 *p)
1254 {
1255 struct rpc_cred *cred = task->tk_rqstp->rq_cred;
1256 struct gss_cl_ctx *ctx = gss_cred_get_ctx(cred);
1257 __be32 seq;
1258 struct kvec iov;
1259 struct xdr_buf verf_buf;
1260 struct xdr_netobj mic;
1261 u32 flav,len;
1262 u32 maj_stat;
1263
1264 dprintk("RPC: %5u %s\n", task->tk_pid, __func__);
1265
1266 flav = ntohl(*p++);
1267 if ((len = ntohl(*p++)) > RPC_MAX_AUTH_SIZE)
1268 goto out_bad;
1269 if (flav != RPC_AUTH_GSS)
1270 goto out_bad;
1271 seq = htonl(task->tk_rqstp->rq_seqno);
1272 iov.iov_base = &seq;
1273 iov.iov_len = sizeof(seq);
1274 xdr_buf_from_iov(&iov, &verf_buf);
1275 mic.data = (u8 *)p;
1276 mic.len = len;
1277
1278 maj_stat = gss_verify_mic(ctx->gc_gss_ctx, &verf_buf, &mic);
1279 if (maj_stat == GSS_S_CONTEXT_EXPIRED)
1280 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1281 if (maj_stat) {
1282 dprintk("RPC: %5u %s: gss_verify_mic returned error 0x%08x\n",
1283 task->tk_pid, __func__, maj_stat);
1284 goto out_bad;
1285 }
1286 /* We leave it to unwrap to calculate au_rslack. For now we just
1287 * calculate the length of the verifier: */
1288 cred->cr_auth->au_verfsize = XDR_QUADLEN(len) + 2;
1289 gss_put_ctx(ctx);
1290 dprintk("RPC: %5u %s: gss_verify_mic succeeded.\n",
1291 task->tk_pid, __func__);
1292 return p + XDR_QUADLEN(len);
1293 out_bad:
1294 gss_put_ctx(ctx);
1295 dprintk("RPC: %5u %s failed.\n", task->tk_pid, __func__);
1296 return NULL;
1297 }
1298
1299 static void gss_wrap_req_encode(kxdreproc_t encode, struct rpc_rqst *rqstp,
1300 __be32 *p, void *obj)
1301 {
1302 struct xdr_stream xdr;
1303
1304 xdr_init_encode(&xdr, &rqstp->rq_snd_buf, p);
1305 encode(rqstp, &xdr, obj);
1306 }
1307
1308 static inline int
1309 gss_wrap_req_integ(struct rpc_cred *cred, struct gss_cl_ctx *ctx,
1310 kxdreproc_t encode, struct rpc_rqst *rqstp,
1311 __be32 *p, void *obj)
1312 {
1313 struct xdr_buf *snd_buf = &rqstp->rq_snd_buf;
1314 struct xdr_buf integ_buf;
1315 __be32 *integ_len = NULL;
1316 struct xdr_netobj mic;
1317 u32 offset;
1318 __be32 *q;
1319 struct kvec *iov;
1320 u32 maj_stat = 0;
1321 int status = -EIO;
1322
1323 integ_len = p++;
1324 offset = (u8 *)p - (u8 *)snd_buf->head[0].iov_base;
1325 *p++ = htonl(rqstp->rq_seqno);
1326
1327 gss_wrap_req_encode(encode, rqstp, p, obj);
1328
1329 if (xdr_buf_subsegment(snd_buf, &integ_buf,
1330 offset, snd_buf->len - offset))
1331 return status;
1332 *integ_len = htonl(integ_buf.len);
1333
1334 /* guess whether we're in the head or the tail: */
1335 if (snd_buf->page_len || snd_buf->tail[0].iov_len)
1336 iov = snd_buf->tail;
1337 else
1338 iov = snd_buf->head;
1339 p = iov->iov_base + iov->iov_len;
1340 mic.data = (u8 *)(p + 1);
1341
1342 maj_stat = gss_get_mic(ctx->gc_gss_ctx, &integ_buf, &mic);
1343 status = -EIO; /* XXX? */
1344 if (maj_stat == GSS_S_CONTEXT_EXPIRED)
1345 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1346 else if (maj_stat)
1347 return status;
1348 q = xdr_encode_opaque(p, NULL, mic.len);
1349
1350 offset = (u8 *)q - (u8 *)p;
1351 iov->iov_len += offset;
1352 snd_buf->len += offset;
1353 return 0;
1354 }
1355
1356 static void
1357 priv_release_snd_buf(struct rpc_rqst *rqstp)
1358 {
1359 int i;
1360
1361 for (i=0; i < rqstp->rq_enc_pages_num; i++)
1362 __free_page(rqstp->rq_enc_pages[i]);
1363 kfree(rqstp->rq_enc_pages);
1364 }
1365
1366 static int
1367 alloc_enc_pages(struct rpc_rqst *rqstp)
1368 {
1369 struct xdr_buf *snd_buf = &rqstp->rq_snd_buf;
1370 int first, last, i;
1371
1372 if (snd_buf->page_len == 0) {
1373 rqstp->rq_enc_pages_num = 0;
1374 return 0;
1375 }
1376
1377 first = snd_buf->page_base >> PAGE_CACHE_SHIFT;
1378 last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_CACHE_SHIFT;
1379 rqstp->rq_enc_pages_num = last - first + 1 + 1;
1380 rqstp->rq_enc_pages
1381 = kmalloc(rqstp->rq_enc_pages_num * sizeof(struct page *),
1382 GFP_NOFS);
1383 if (!rqstp->rq_enc_pages)
1384 goto out;
1385 for (i=0; i < rqstp->rq_enc_pages_num; i++) {
1386 rqstp->rq_enc_pages[i] = alloc_page(GFP_NOFS);
1387 if (rqstp->rq_enc_pages[i] == NULL)
1388 goto out_free;
1389 }
1390 rqstp->rq_release_snd_buf = priv_release_snd_buf;
1391 return 0;
1392 out_free:
1393 rqstp->rq_enc_pages_num = i;
1394 priv_release_snd_buf(rqstp);
1395 out:
1396 return -EAGAIN;
1397 }
1398
1399 static inline int
1400 gss_wrap_req_priv(struct rpc_cred *cred, struct gss_cl_ctx *ctx,
1401 kxdreproc_t encode, struct rpc_rqst *rqstp,
1402 __be32 *p, void *obj)
1403 {
1404 struct xdr_buf *snd_buf = &rqstp->rq_snd_buf;
1405 u32 offset;
1406 u32 maj_stat;
1407 int status;
1408 __be32 *opaque_len;
1409 struct page **inpages;
1410 int first;
1411 int pad;
1412 struct kvec *iov;
1413 char *tmp;
1414
1415 opaque_len = p++;
1416 offset = (u8 *)p - (u8 *)snd_buf->head[0].iov_base;
1417 *p++ = htonl(rqstp->rq_seqno);
1418
1419 gss_wrap_req_encode(encode, rqstp, p, obj);
1420
1421 status = alloc_enc_pages(rqstp);
1422 if (status)
1423 return status;
1424 first = snd_buf->page_base >> PAGE_CACHE_SHIFT;
1425 inpages = snd_buf->pages + first;
1426 snd_buf->pages = rqstp->rq_enc_pages;
1427 snd_buf->page_base -= first << PAGE_CACHE_SHIFT;
1428 /*
1429 * Give the tail its own page, in case we need extra space in the
1430 * head when wrapping:
1431 *
1432 * call_allocate() allocates twice the slack space required
1433 * by the authentication flavor to rq_callsize.
1434 * For GSS, slack is GSS_CRED_SLACK.
1435 */
1436 if (snd_buf->page_len || snd_buf->tail[0].iov_len) {
1437 tmp = page_address(rqstp->rq_enc_pages[rqstp->rq_enc_pages_num - 1]);
1438 memcpy(tmp, snd_buf->tail[0].iov_base, snd_buf->tail[0].iov_len);
1439 snd_buf->tail[0].iov_base = tmp;
1440 }
1441 maj_stat = gss_wrap(ctx->gc_gss_ctx, offset, snd_buf, inpages);
1442 /* slack space should prevent this ever happening: */
1443 BUG_ON(snd_buf->len > snd_buf->buflen);
1444 status = -EIO;
1445 /* We're assuming that when GSS_S_CONTEXT_EXPIRED, the encryption was
1446 * done anyway, so it's safe to put the request on the wire: */
1447 if (maj_stat == GSS_S_CONTEXT_EXPIRED)
1448 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1449 else if (maj_stat)
1450 return status;
1451
1452 *opaque_len = htonl(snd_buf->len - offset);
1453 /* guess whether we're in the head or the tail: */
1454 if (snd_buf->page_len || snd_buf->tail[0].iov_len)
1455 iov = snd_buf->tail;
1456 else
1457 iov = snd_buf->head;
1458 p = iov->iov_base + iov->iov_len;
1459 pad = 3 - ((snd_buf->len - offset - 1) & 3);
1460 memset(p, 0, pad);
1461 iov->iov_len += pad;
1462 snd_buf->len += pad;
1463
1464 return 0;
1465 }
1466
1467 static int
1468 gss_wrap_req(struct rpc_task *task,
1469 kxdreproc_t encode, void *rqstp, __be32 *p, void *obj)
1470 {
1471 struct rpc_cred *cred = task->tk_rqstp->rq_cred;
1472 struct gss_cred *gss_cred = container_of(cred, struct gss_cred,
1473 gc_base);
1474 struct gss_cl_ctx *ctx = gss_cred_get_ctx(cred);
1475 int status = -EIO;
1476
1477 dprintk("RPC: %5u %s\n", task->tk_pid, __func__);
1478 if (ctx->gc_proc != RPC_GSS_PROC_DATA) {
1479 /* The spec seems a little ambiguous here, but I think that not
1480 * wrapping context destruction requests makes the most sense.
1481 */
1482 gss_wrap_req_encode(encode, rqstp, p, obj);
1483 status = 0;
1484 goto out;
1485 }
1486 switch (gss_cred->gc_service) {
1487 case RPC_GSS_SVC_NONE:
1488 gss_wrap_req_encode(encode, rqstp, p, obj);
1489 status = 0;
1490 break;
1491 case RPC_GSS_SVC_INTEGRITY:
1492 status = gss_wrap_req_integ(cred, ctx, encode, rqstp, p, obj);
1493 break;
1494 case RPC_GSS_SVC_PRIVACY:
1495 status = gss_wrap_req_priv(cred, ctx, encode, rqstp, p, obj);
1496 break;
1497 }
1498 out:
1499 gss_put_ctx(ctx);
1500 dprintk("RPC: %5u %s returning %d\n", task->tk_pid, __func__, status);
1501 return status;
1502 }
1503
1504 static inline int
1505 gss_unwrap_resp_integ(struct rpc_cred *cred, struct gss_cl_ctx *ctx,
1506 struct rpc_rqst *rqstp, __be32 **p)
1507 {
1508 struct xdr_buf *rcv_buf = &rqstp->rq_rcv_buf;
1509 struct xdr_buf integ_buf;
1510 struct xdr_netobj mic;
1511 u32 data_offset, mic_offset;
1512 u32 integ_len;
1513 u32 maj_stat;
1514 int status = -EIO;
1515
1516 integ_len = ntohl(*(*p)++);
1517 if (integ_len & 3)
1518 return status;
1519 data_offset = (u8 *)(*p) - (u8 *)rcv_buf->head[0].iov_base;
1520 mic_offset = integ_len + data_offset;
1521 if (mic_offset > rcv_buf->len)
1522 return status;
1523 if (ntohl(*(*p)++) != rqstp->rq_seqno)
1524 return status;
1525
1526 if (xdr_buf_subsegment(rcv_buf, &integ_buf, data_offset,
1527 mic_offset - data_offset))
1528 return status;
1529
1530 if (xdr_buf_read_netobj(rcv_buf, &mic, mic_offset))
1531 return status;
1532
1533 maj_stat = gss_verify_mic(ctx->gc_gss_ctx, &integ_buf, &mic);
1534 if (maj_stat == GSS_S_CONTEXT_EXPIRED)
1535 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1536 if (maj_stat != GSS_S_COMPLETE)
1537 return status;
1538 return 0;
1539 }
1540
1541 static inline int
1542 gss_unwrap_resp_priv(struct rpc_cred *cred, struct gss_cl_ctx *ctx,
1543 struct rpc_rqst *rqstp, __be32 **p)
1544 {
1545 struct xdr_buf *rcv_buf = &rqstp->rq_rcv_buf;
1546 u32 offset;
1547 u32 opaque_len;
1548 u32 maj_stat;
1549 int status = -EIO;
1550
1551 opaque_len = ntohl(*(*p)++);
1552 offset = (u8 *)(*p) - (u8 *)rcv_buf->head[0].iov_base;
1553 if (offset + opaque_len > rcv_buf->len)
1554 return status;
1555 /* remove padding: */
1556 rcv_buf->len = offset + opaque_len;
1557
1558 maj_stat = gss_unwrap(ctx->gc_gss_ctx, offset, rcv_buf);
1559 if (maj_stat == GSS_S_CONTEXT_EXPIRED)
1560 clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags);
1561 if (maj_stat != GSS_S_COMPLETE)
1562 return status;
1563 if (ntohl(*(*p)++) != rqstp->rq_seqno)
1564 return status;
1565
1566 return 0;
1567 }
1568
1569 static int
1570 gss_unwrap_req_decode(kxdrdproc_t decode, struct rpc_rqst *rqstp,
1571 __be32 *p, void *obj)
1572 {
1573 struct xdr_stream xdr;
1574
1575 xdr_init_decode(&xdr, &rqstp->rq_rcv_buf, p);
1576 return decode(rqstp, &xdr, obj);
1577 }
1578
1579 static int
1580 gss_unwrap_resp(struct rpc_task *task,
1581 kxdrdproc_t decode, void *rqstp, __be32 *p, void *obj)
1582 {
1583 struct rpc_cred *cred = task->tk_rqstp->rq_cred;
1584 struct gss_cred *gss_cred = container_of(cred, struct gss_cred,
1585 gc_base);
1586 struct gss_cl_ctx *ctx = gss_cred_get_ctx(cred);
1587 __be32 *savedp = p;
1588 struct kvec *head = ((struct rpc_rqst *)rqstp)->rq_rcv_buf.head;
1589 int savedlen = head->iov_len;
1590 int status = -EIO;
1591
1592 if (ctx->gc_proc != RPC_GSS_PROC_DATA)
1593 goto out_decode;
1594 switch (gss_cred->gc_service) {
1595 case RPC_GSS_SVC_NONE:
1596 break;
1597 case RPC_GSS_SVC_INTEGRITY:
1598 status = gss_unwrap_resp_integ(cred, ctx, rqstp, &p);
1599 if (status)
1600 goto out;
1601 break;
1602 case RPC_GSS_SVC_PRIVACY:
1603 status = gss_unwrap_resp_priv(cred, ctx, rqstp, &p);
1604 if (status)
1605 goto out;
1606 break;
1607 }
1608 /* take into account extra slack for integrity and privacy cases: */
1609 cred->cr_auth->au_rslack = cred->cr_auth->au_verfsize + (p - savedp)
1610 + (savedlen - head->iov_len);
1611 out_decode:
1612 status = gss_unwrap_req_decode(decode, rqstp, p, obj);
1613 out:
1614 gss_put_ctx(ctx);
1615 dprintk("RPC: %5u %s returning %d\n",
1616 task->tk_pid, __func__, status);
1617 return status;
1618 }
1619
1620 static const struct rpc_authops authgss_ops = {
1621 .owner = THIS_MODULE,
1622 .au_flavor = RPC_AUTH_GSS,
1623 .au_name = "RPCSEC_GSS",
1624 .create = gss_create,
1625 .destroy = gss_destroy,
1626 .lookup_cred = gss_lookup_cred,
1627 .crcreate = gss_create_cred,
1628 .pipes_create = gss_pipes_dentries_create,
1629 .pipes_destroy = gss_pipes_dentries_destroy,
1630 .list_pseudoflavors = gss_mech_list_pseudoflavors,
1631 };
1632
1633 static const struct rpc_credops gss_credops = {
1634 .cr_name = "AUTH_GSS",
1635 .crdestroy = gss_destroy_cred,
1636 .cr_init = gss_cred_init,
1637 .crbind = rpcauth_generic_bind_cred,
1638 .crmatch = gss_match,
1639 .crmarshal = gss_marshal,
1640 .crrefresh = gss_refresh,
1641 .crvalidate = gss_validate,
1642 .crwrap_req = gss_wrap_req,
1643 .crunwrap_resp = gss_unwrap_resp,
1644 };
1645
1646 static const struct rpc_credops gss_nullops = {
1647 .cr_name = "AUTH_GSS",
1648 .crdestroy = gss_destroy_nullcred,
1649 .crbind = rpcauth_generic_bind_cred,
1650 .crmatch = gss_match,
1651 .crmarshal = gss_marshal,
1652 .crrefresh = gss_refresh_null,
1653 .crvalidate = gss_validate,
1654 .crwrap_req = gss_wrap_req,
1655 .crunwrap_resp = gss_unwrap_resp,
1656 };
1657
1658 static const struct rpc_pipe_ops gss_upcall_ops_v0 = {
1659 .upcall = rpc_pipe_generic_upcall,
1660 .downcall = gss_pipe_downcall,
1661 .destroy_msg = gss_pipe_destroy_msg,
1662 .open_pipe = gss_pipe_open_v0,
1663 .release_pipe = gss_pipe_release,
1664 };
1665
1666 static const struct rpc_pipe_ops gss_upcall_ops_v1 = {
1667 .upcall = rpc_pipe_generic_upcall,
1668 .downcall = gss_pipe_downcall,
1669 .destroy_msg = gss_pipe_destroy_msg,
1670 .open_pipe = gss_pipe_open_v1,
1671 .release_pipe = gss_pipe_release,
1672 };
1673
1674 static __net_init int rpcsec_gss_init_net(struct net *net)
1675 {
1676 return gss_svc_init_net(net);
1677 }
1678
1679 static __net_exit void rpcsec_gss_exit_net(struct net *net)
1680 {
1681 gss_svc_shutdown_net(net);
1682 }
1683
1684 static struct pernet_operations rpcsec_gss_net_ops = {
1685 .init = rpcsec_gss_init_net,
1686 .exit = rpcsec_gss_exit_net,
1687 };
1688
1689 /*
1690 * Initialize RPCSEC_GSS module
1691 */
1692 static int __init init_rpcsec_gss(void)
1693 {
1694 int err = 0;
1695
1696 err = rpcauth_register(&authgss_ops);
1697 if (err)
1698 goto out;
1699 err = gss_svc_init();
1700 if (err)
1701 goto out_unregister;
1702 err = register_pernet_subsys(&rpcsec_gss_net_ops);
1703 if (err)
1704 goto out_svc_exit;
1705 rpc_init_wait_queue(&pipe_version_rpc_waitqueue, "gss pipe version");
1706 return 0;
1707 out_svc_exit:
1708 gss_svc_shutdown();
1709 out_unregister:
1710 rpcauth_unregister(&authgss_ops);
1711 out:
1712 return err;
1713 }
1714
1715 static void __exit exit_rpcsec_gss(void)
1716 {
1717 unregister_pernet_subsys(&rpcsec_gss_net_ops);
1718 gss_svc_shutdown();
1719 rpcauth_unregister(&authgss_ops);
1720 rcu_barrier(); /* Wait for completion of call_rcu()'s */
1721 }
1722
1723 MODULE_LICENSE("GPL");
1724 module_param_named(expired_cred_retry_delay,
1725 gss_expired_cred_retry_delay,
1726 uint, 0644);
1727 MODULE_PARM_DESC(expired_cred_retry_delay, "Timeout (in seconds) until "
1728 "the RPC engine retries an expired credential");
1729
1730 module_init(init_rpcsec_gss)
1731 module_exit(exit_rpcsec_gss)