PageRenderTime 87ms CodeModel.GetById 22ms app.highlight 53ms RepoModel.GetById 0ms app.codeStats 1ms

/fs/gfs2/dir.c

https://bitbucket.org/emiliolopez/linux
C | 2197 lines | 1569 code | 313 blank | 315 comment | 284 complexity | 98b4814a62280bdfb83b20b9a9d7b763 MD5 | raw file
Possible License(s): GPL-2.0, LGPL-2.0, AGPL-1.0

Large files files are truncated, but you can click here to view the full file

   1/*
   2 * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
   3 * Copyright (C) 2004-2006 Red Hat, Inc.  All rights reserved.
   4 *
   5 * This copyrighted material is made available to anyone wishing to use,
   6 * modify, copy, or redistribute it subject to the terms and conditions
   7 * of the GNU General Public License version 2.
   8 */
   9
  10/*
  11 * Implements Extendible Hashing as described in:
  12 *   "Extendible Hashing" by Fagin, et al in
  13 *     __ACM Trans. on Database Systems__, Sept 1979.
  14 *
  15 *
  16 * Here's the layout of dirents which is essentially the same as that of ext2
  17 * within a single block. The field de_name_len is the number of bytes
  18 * actually required for the name (no null terminator). The field de_rec_len
  19 * is the number of bytes allocated to the dirent. The offset of the next
  20 * dirent in the block is (dirent + dirent->de_rec_len). When a dirent is
  21 * deleted, the preceding dirent inherits its allocated space, ie
  22 * prev->de_rec_len += deleted->de_rec_len. Since the next dirent is obtained
  23 * by adding de_rec_len to the current dirent, this essentially causes the
  24 * deleted dirent to get jumped over when iterating through all the dirents.
  25 *
  26 * When deleting the first dirent in a block, there is no previous dirent so
  27 * the field de_ino is set to zero to designate it as deleted. When allocating
  28 * a dirent, gfs2_dirent_alloc iterates through the dirents in a block. If the
  29 * first dirent has (de_ino == 0) and de_rec_len is large enough, this first
  30 * dirent is allocated. Otherwise it must go through all the 'used' dirents
  31 * searching for one in which the amount of total space minus the amount of
  32 * used space will provide enough space for the new dirent.
  33 *
  34 * There are two types of blocks in which dirents reside. In a stuffed dinode,
  35 * the dirents begin at offset sizeof(struct gfs2_dinode) from the beginning of
  36 * the block.  In leaves, they begin at offset sizeof(struct gfs2_leaf) from the
  37 * beginning of the leaf block. The dirents reside in leaves when
  38 *
  39 * dip->i_diskflags & GFS2_DIF_EXHASH is true
  40 *
  41 * Otherwise, the dirents are "linear", within a single stuffed dinode block.
  42 *
  43 * When the dirents are in leaves, the actual contents of the directory file are
  44 * used as an array of 64-bit block pointers pointing to the leaf blocks. The
  45 * dirents are NOT in the directory file itself. There can be more than one
  46 * block pointer in the array that points to the same leaf. In fact, when a
  47 * directory is first converted from linear to exhash, all of the pointers
  48 * point to the same leaf.
  49 *
  50 * When a leaf is completely full, the size of the hash table can be
  51 * doubled unless it is already at the maximum size which is hard coded into
  52 * GFS2_DIR_MAX_DEPTH. After that, leaves are chained together in a linked list,
  53 * but never before the maximum hash table size has been reached.
  54 */
  55
  56#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  57
  58#include <linux/slab.h>
  59#include <linux/spinlock.h>
  60#include <linux/buffer_head.h>
  61#include <linux/sort.h>
  62#include <linux/gfs2_ondisk.h>
  63#include <linux/crc32.h>
  64#include <linux/vmalloc.h>
  65#include <linux/bio.h>
  66
  67#include "gfs2.h"
  68#include "incore.h"
  69#include "dir.h"
  70#include "glock.h"
  71#include "inode.h"
  72#include "meta_io.h"
  73#include "quota.h"
  74#include "rgrp.h"
  75#include "trans.h"
  76#include "bmap.h"
  77#include "util.h"
  78
  79#define IS_LEAF     1 /* Hashed (leaf) directory */
  80#define IS_DINODE   2 /* Linear (stuffed dinode block) directory */
  81
  82#define MAX_RA_BLOCKS 32 /* max read-ahead blocks */
  83
  84#define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1)
  85#define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1))
  86#define GFS2_HASH_INDEX_MASK 0xffffc000
  87#define GFS2_USE_HASH_FLAG 0x2000
  88
  89struct qstr gfs2_qdot __read_mostly;
  90struct qstr gfs2_qdotdot __read_mostly;
  91
  92typedef int (*gfs2_dscan_t)(const struct gfs2_dirent *dent,
  93			    const struct qstr *name, void *opaque);
  94
  95int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
  96			    struct buffer_head **bhp)
  97{
  98	struct buffer_head *bh;
  99
 100	bh = gfs2_meta_new(ip->i_gl, block);
 101	gfs2_trans_add_meta(ip->i_gl, bh);
 102	gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD);
 103	gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header));
 104	*bhp = bh;
 105	return 0;
 106}
 107
 108static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block,
 109					struct buffer_head **bhp)
 110{
 111	struct buffer_head *bh;
 112	int error;
 113
 114	error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh);
 115	if (error)
 116		return error;
 117	if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) {
 118		brelse(bh);
 119		return -EIO;
 120	}
 121	*bhp = bh;
 122	return 0;
 123}
 124
 125static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf,
 126				  unsigned int offset, unsigned int size)
 127{
 128	struct buffer_head *dibh;
 129	int error;
 130
 131	error = gfs2_meta_inode_buffer(ip, &dibh);
 132	if (error)
 133		return error;
 134
 135	gfs2_trans_add_meta(ip->i_gl, dibh);
 136	memcpy(dibh->b_data + offset + sizeof(struct gfs2_dinode), buf, size);
 137	if (ip->i_inode.i_size < offset + size)
 138		i_size_write(&ip->i_inode, offset + size);
 139	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
 140	gfs2_dinode_out(ip, dibh->b_data);
 141
 142	brelse(dibh);
 143
 144	return size;
 145}
 146
 147
 148
 149/**
 150 * gfs2_dir_write_data - Write directory information to the inode
 151 * @ip: The GFS2 inode
 152 * @buf: The buffer containing information to be written
 153 * @offset: The file offset to start writing at
 154 * @size: The amount of data to write
 155 *
 156 * Returns: The number of bytes correctly written or error code
 157 */
 158static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf,
 159			       u64 offset, unsigned int size)
 160{
 161	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
 162	struct buffer_head *dibh;
 163	u64 lblock, dblock;
 164	u32 extlen = 0;
 165	unsigned int o;
 166	int copied = 0;
 167	int error = 0;
 168	int new = 0;
 169
 170	if (!size)
 171		return 0;
 172
 173	if (gfs2_is_stuffed(ip) &&
 174	    offset + size <= sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode))
 175		return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset,
 176					      size);
 177
 178	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
 179		return -EINVAL;
 180
 181	if (gfs2_is_stuffed(ip)) {
 182		error = gfs2_unstuff_dinode(ip, NULL);
 183		if (error)
 184			return error;
 185	}
 186
 187	lblock = offset;
 188	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
 189
 190	while (copied < size) {
 191		unsigned int amount;
 192		struct buffer_head *bh;
 193
 194		amount = size - copied;
 195		if (amount > sdp->sd_sb.sb_bsize - o)
 196			amount = sdp->sd_sb.sb_bsize - o;
 197
 198		if (!extlen) {
 199			new = 1;
 200			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
 201						&dblock, &extlen);
 202			if (error)
 203				goto fail;
 204			error = -EIO;
 205			if (gfs2_assert_withdraw(sdp, dblock))
 206				goto fail;
 207		}
 208
 209		if (amount == sdp->sd_jbsize || new)
 210			error = gfs2_dir_get_new_buffer(ip, dblock, &bh);
 211		else
 212			error = gfs2_dir_get_existing_buffer(ip, dblock, &bh);
 213
 214		if (error)
 215			goto fail;
 216
 217		gfs2_trans_add_meta(ip->i_gl, bh);
 218		memcpy(bh->b_data + o, buf, amount);
 219		brelse(bh);
 220
 221		buf += amount;
 222		copied += amount;
 223		lblock++;
 224		dblock++;
 225		extlen--;
 226
 227		o = sizeof(struct gfs2_meta_header);
 228	}
 229
 230out:
 231	error = gfs2_meta_inode_buffer(ip, &dibh);
 232	if (error)
 233		return error;
 234
 235	if (ip->i_inode.i_size < offset + copied)
 236		i_size_write(&ip->i_inode, offset + copied);
 237	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
 238
 239	gfs2_trans_add_meta(ip->i_gl, dibh);
 240	gfs2_dinode_out(ip, dibh->b_data);
 241	brelse(dibh);
 242
 243	return copied;
 244fail:
 245	if (copied)
 246		goto out;
 247	return error;
 248}
 249
 250static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf,
 251				 unsigned int size)
 252{
 253	struct buffer_head *dibh;
 254	int error;
 255
 256	error = gfs2_meta_inode_buffer(ip, &dibh);
 257	if (!error) {
 258		memcpy(buf, dibh->b_data + sizeof(struct gfs2_dinode), size);
 259		brelse(dibh);
 260	}
 261
 262	return (error) ? error : size;
 263}
 264
 265
 266/**
 267 * gfs2_dir_read_data - Read a data from a directory inode
 268 * @ip: The GFS2 Inode
 269 * @buf: The buffer to place result into
 270 * @size: Amount of data to transfer
 271 *
 272 * Returns: The amount of data actually copied or the error
 273 */
 274static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf,
 275			      unsigned int size)
 276{
 277	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
 278	u64 lblock, dblock;
 279	u32 extlen = 0;
 280	unsigned int o;
 281	int copied = 0;
 282	int error = 0;
 283
 284	if (gfs2_is_stuffed(ip))
 285		return gfs2_dir_read_stuffed(ip, buf, size);
 286
 287	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
 288		return -EINVAL;
 289
 290	lblock = 0;
 291	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
 292
 293	while (copied < size) {
 294		unsigned int amount;
 295		struct buffer_head *bh;
 296		int new;
 297
 298		amount = size - copied;
 299		if (amount > sdp->sd_sb.sb_bsize - o)
 300			amount = sdp->sd_sb.sb_bsize - o;
 301
 302		if (!extlen) {
 303			new = 0;
 304			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
 305						&dblock, &extlen);
 306			if (error || !dblock)
 307				goto fail;
 308			BUG_ON(extlen < 1);
 309			bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
 310		} else {
 311			error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh);
 312			if (error)
 313				goto fail;
 314		}
 315		error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD);
 316		if (error) {
 317			brelse(bh);
 318			goto fail;
 319		}
 320		dblock++;
 321		extlen--;
 322		memcpy(buf, bh->b_data + o, amount);
 323		brelse(bh);
 324		buf += (amount/sizeof(__be64));
 325		copied += amount;
 326		lblock++;
 327		o = sizeof(struct gfs2_meta_header);
 328	}
 329
 330	return copied;
 331fail:
 332	return (copied) ? copied : error;
 333}
 334
 335/**
 336 * gfs2_dir_get_hash_table - Get pointer to the dir hash table
 337 * @ip: The inode in question
 338 *
 339 * Returns: The hash table or an error
 340 */
 341
 342static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip)
 343{
 344	struct inode *inode = &ip->i_inode;
 345	int ret;
 346	u32 hsize;
 347	__be64 *hc;
 348
 349	BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH));
 350
 351	hc = ip->i_hash_cache;
 352	if (hc)
 353		return hc;
 354
 355	hsize = BIT(ip->i_depth);
 356	hsize *= sizeof(__be64);
 357	if (hsize != i_size_read(&ip->i_inode)) {
 358		gfs2_consist_inode(ip);
 359		return ERR_PTR(-EIO);
 360	}
 361
 362	hc = kmalloc(hsize, GFP_NOFS | __GFP_NOWARN);
 363	if (hc == NULL)
 364		hc = __vmalloc(hsize, GFP_NOFS, PAGE_KERNEL);
 365
 366	if (hc == NULL)
 367		return ERR_PTR(-ENOMEM);
 368
 369	ret = gfs2_dir_read_data(ip, hc, hsize);
 370	if (ret < 0) {
 371		kvfree(hc);
 372		return ERR_PTR(ret);
 373	}
 374
 375	spin_lock(&inode->i_lock);
 376	if (likely(!ip->i_hash_cache)) {
 377		ip->i_hash_cache = hc;
 378		hc = NULL;
 379	}
 380	spin_unlock(&inode->i_lock);
 381	kvfree(hc);
 382
 383	return ip->i_hash_cache;
 384}
 385
 386/**
 387 * gfs2_dir_hash_inval - Invalidate dir hash
 388 * @ip: The directory inode
 389 *
 390 * Must be called with an exclusive glock, or during glock invalidation.
 391 */
 392void gfs2_dir_hash_inval(struct gfs2_inode *ip)
 393{
 394	__be64 *hc;
 395
 396	spin_lock(&ip->i_inode.i_lock);
 397	hc = ip->i_hash_cache;
 398	ip->i_hash_cache = NULL;
 399	spin_unlock(&ip->i_inode.i_lock);
 400
 401	kvfree(hc);
 402}
 403
 404static inline int gfs2_dirent_sentinel(const struct gfs2_dirent *dent)
 405{
 406	return dent->de_inum.no_addr == 0 || dent->de_inum.no_formal_ino == 0;
 407}
 408
 409static inline int __gfs2_dirent_find(const struct gfs2_dirent *dent,
 410				     const struct qstr *name, int ret)
 411{
 412	if (!gfs2_dirent_sentinel(dent) &&
 413	    be32_to_cpu(dent->de_hash) == name->hash &&
 414	    be16_to_cpu(dent->de_name_len) == name->len &&
 415	    memcmp(dent+1, name->name, name->len) == 0)
 416		return ret;
 417	return 0;
 418}
 419
 420static int gfs2_dirent_find(const struct gfs2_dirent *dent,
 421			    const struct qstr *name,
 422			    void *opaque)
 423{
 424	return __gfs2_dirent_find(dent, name, 1);
 425}
 426
 427static int gfs2_dirent_prev(const struct gfs2_dirent *dent,
 428			    const struct qstr *name,
 429			    void *opaque)
 430{
 431	return __gfs2_dirent_find(dent, name, 2);
 432}
 433
 434/*
 435 * name->name holds ptr to start of block.
 436 * name->len holds size of block.
 437 */
 438static int gfs2_dirent_last(const struct gfs2_dirent *dent,
 439			    const struct qstr *name,
 440			    void *opaque)
 441{
 442	const char *start = name->name;
 443	const char *end = (const char *)dent + be16_to_cpu(dent->de_rec_len);
 444	if (name->len == (end - start))
 445		return 1;
 446	return 0;
 447}
 448
 449/* Look for the dirent that contains the offset specified in data. Once we
 450 * find that dirent, there must be space available there for the new dirent */
 451static int gfs2_dirent_find_offset(const struct gfs2_dirent *dent,
 452				  const struct qstr *name,
 453				  void *ptr)
 454{
 455	unsigned required = GFS2_DIRENT_SIZE(name->len);
 456	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
 457	unsigned totlen = be16_to_cpu(dent->de_rec_len);
 458
 459	if (ptr < (void *)dent || ptr >= (void *)dent + totlen)
 460		return 0;
 461	if (gfs2_dirent_sentinel(dent))
 462		actual = 0;
 463	if (ptr < (void *)dent + actual)
 464		return -1;
 465	if ((void *)dent + totlen >= ptr + required)
 466		return 1;
 467	return -1;
 468}
 469
 470static int gfs2_dirent_find_space(const struct gfs2_dirent *dent,
 471				  const struct qstr *name,
 472				  void *opaque)
 473{
 474	unsigned required = GFS2_DIRENT_SIZE(name->len);
 475	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
 476	unsigned totlen = be16_to_cpu(dent->de_rec_len);
 477
 478	if (gfs2_dirent_sentinel(dent))
 479		actual = 0;
 480	if (totlen - actual >= required)
 481		return 1;
 482	return 0;
 483}
 484
 485struct dirent_gather {
 486	const struct gfs2_dirent **pdent;
 487	unsigned offset;
 488};
 489
 490static int gfs2_dirent_gather(const struct gfs2_dirent *dent,
 491			      const struct qstr *name,
 492			      void *opaque)
 493{
 494	struct dirent_gather *g = opaque;
 495	if (!gfs2_dirent_sentinel(dent)) {
 496		g->pdent[g->offset++] = dent;
 497	}
 498	return 0;
 499}
 500
 501/*
 502 * Other possible things to check:
 503 * - Inode located within filesystem size (and on valid block)
 504 * - Valid directory entry type
 505 * Not sure how heavy-weight we want to make this... could also check
 506 * hash is correct for example, but that would take a lot of extra time.
 507 * For now the most important thing is to check that the various sizes
 508 * are correct.
 509 */
 510static int gfs2_check_dirent(struct gfs2_dirent *dent, unsigned int offset,
 511			     unsigned int size, unsigned int len, int first)
 512{
 513	const char *msg = "gfs2_dirent too small";
 514	if (unlikely(size < sizeof(struct gfs2_dirent)))
 515		goto error;
 516	msg = "gfs2_dirent misaligned";
 517	if (unlikely(offset & 0x7))
 518		goto error;
 519	msg = "gfs2_dirent points beyond end of block";
 520	if (unlikely(offset + size > len))
 521		goto error;
 522	msg = "zero inode number";
 523	if (unlikely(!first && gfs2_dirent_sentinel(dent)))
 524		goto error;
 525	msg = "name length is greater than space in dirent";
 526	if (!gfs2_dirent_sentinel(dent) &&
 527	    unlikely(sizeof(struct gfs2_dirent)+be16_to_cpu(dent->de_name_len) >
 528		     size))
 529		goto error;
 530	return 0;
 531error:
 532	pr_warn("%s: %s (%s)\n",
 533		__func__, msg, first ? "first in block" : "not first in block");
 534	return -EIO;
 535}
 536
 537static int gfs2_dirent_offset(const void *buf)
 538{
 539	const struct gfs2_meta_header *h = buf;
 540	int offset;
 541
 542	BUG_ON(buf == NULL);
 543
 544	switch(be32_to_cpu(h->mh_type)) {
 545	case GFS2_METATYPE_LF:
 546		offset = sizeof(struct gfs2_leaf);
 547		break;
 548	case GFS2_METATYPE_DI:
 549		offset = sizeof(struct gfs2_dinode);
 550		break;
 551	default:
 552		goto wrong_type;
 553	}
 554	return offset;
 555wrong_type:
 556	pr_warn("%s: wrong block type %u\n", __func__, be32_to_cpu(h->mh_type));
 557	return -1;
 558}
 559
 560static struct gfs2_dirent *gfs2_dirent_scan(struct inode *inode, void *buf,
 561					    unsigned int len, gfs2_dscan_t scan,
 562					    const struct qstr *name,
 563					    void *opaque)
 564{
 565	struct gfs2_dirent *dent, *prev;
 566	unsigned offset;
 567	unsigned size;
 568	int ret = 0;
 569
 570	ret = gfs2_dirent_offset(buf);
 571	if (ret < 0)
 572		goto consist_inode;
 573
 574	offset = ret;
 575	prev = NULL;
 576	dent = buf + offset;
 577	size = be16_to_cpu(dent->de_rec_len);
 578	if (gfs2_check_dirent(dent, offset, size, len, 1))
 579		goto consist_inode;
 580	do {
 581		ret = scan(dent, name, opaque);
 582		if (ret)
 583			break;
 584		offset += size;
 585		if (offset == len)
 586			break;
 587		prev = dent;
 588		dent = buf + offset;
 589		size = be16_to_cpu(dent->de_rec_len);
 590		if (gfs2_check_dirent(dent, offset, size, len, 0))
 591			goto consist_inode;
 592	} while(1);
 593
 594	switch(ret) {
 595	case 0:
 596		return NULL;
 597	case 1:
 598		return dent;
 599	case 2:
 600		return prev ? prev : dent;
 601	default:
 602		BUG_ON(ret > 0);
 603		return ERR_PTR(ret);
 604	}
 605
 606consist_inode:
 607	gfs2_consist_inode(GFS2_I(inode));
 608	return ERR_PTR(-EIO);
 609}
 610
 611static int dirent_check_reclen(struct gfs2_inode *dip,
 612			       const struct gfs2_dirent *d, const void *end_p)
 613{
 614	const void *ptr = d;
 615	u16 rec_len = be16_to_cpu(d->de_rec_len);
 616
 617	if (unlikely(rec_len < sizeof(struct gfs2_dirent)))
 618		goto broken;
 619	ptr += rec_len;
 620	if (ptr < end_p)
 621		return rec_len;
 622	if (ptr == end_p)
 623		return -ENOENT;
 624broken:
 625	gfs2_consist_inode(dip);
 626	return -EIO;
 627}
 628
 629/**
 630 * dirent_next - Next dirent
 631 * @dip: the directory
 632 * @bh: The buffer
 633 * @dent: Pointer to list of dirents
 634 *
 635 * Returns: 0 on success, error code otherwise
 636 */
 637
 638static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh,
 639		       struct gfs2_dirent **dent)
 640{
 641	struct gfs2_dirent *cur = *dent, *tmp;
 642	char *bh_end = bh->b_data + bh->b_size;
 643	int ret;
 644
 645	ret = dirent_check_reclen(dip, cur, bh_end);
 646	if (ret < 0)
 647		return ret;
 648
 649	tmp = (void *)cur + ret;
 650	ret = dirent_check_reclen(dip, tmp, bh_end);
 651	if (ret == -EIO)
 652		return ret;
 653
 654        /* Only the first dent could ever have de_inum.no_addr == 0 */
 655	if (gfs2_dirent_sentinel(tmp)) {
 656		gfs2_consist_inode(dip);
 657		return -EIO;
 658	}
 659
 660	*dent = tmp;
 661	return 0;
 662}
 663
 664/**
 665 * dirent_del - Delete a dirent
 666 * @dip: The GFS2 inode
 667 * @bh: The buffer
 668 * @prev: The previous dirent
 669 * @cur: The current dirent
 670 *
 671 */
 672
 673static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh,
 674		       struct gfs2_dirent *prev, struct gfs2_dirent *cur)
 675{
 676	u16 cur_rec_len, prev_rec_len;
 677
 678	if (gfs2_dirent_sentinel(cur)) {
 679		gfs2_consist_inode(dip);
 680		return;
 681	}
 682
 683	gfs2_trans_add_meta(dip->i_gl, bh);
 684
 685	/* If there is no prev entry, this is the first entry in the block.
 686	   The de_rec_len is already as big as it needs to be.  Just zero
 687	   out the inode number and return.  */
 688
 689	if (!prev) {
 690		cur->de_inum.no_addr = 0;
 691		cur->de_inum.no_formal_ino = 0;
 692		return;
 693	}
 694
 695	/*  Combine this dentry with the previous one.  */
 696
 697	prev_rec_len = be16_to_cpu(prev->de_rec_len);
 698	cur_rec_len = be16_to_cpu(cur->de_rec_len);
 699
 700	if ((char *)prev + prev_rec_len != (char *)cur)
 701		gfs2_consist_inode(dip);
 702	if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size)
 703		gfs2_consist_inode(dip);
 704
 705	prev_rec_len += cur_rec_len;
 706	prev->de_rec_len = cpu_to_be16(prev_rec_len);
 707}
 708
 709
 710static struct gfs2_dirent *do_init_dirent(struct inode *inode,
 711					  struct gfs2_dirent *dent,
 712					  const struct qstr *name,
 713					  struct buffer_head *bh,
 714					  unsigned offset)
 715{
 716	struct gfs2_inode *ip = GFS2_I(inode);
 717	struct gfs2_dirent *ndent;
 718	unsigned totlen;
 719
 720	totlen = be16_to_cpu(dent->de_rec_len);
 721	BUG_ON(offset + name->len > totlen);
 722	gfs2_trans_add_meta(ip->i_gl, bh);
 723	ndent = (struct gfs2_dirent *)((char *)dent + offset);
 724	dent->de_rec_len = cpu_to_be16(offset);
 725	gfs2_qstr2dirent(name, totlen - offset, ndent);
 726	return ndent;
 727}
 728
 729
 730/*
 731 * Takes a dent from which to grab space as an argument. Returns the
 732 * newly created dent.
 733 */
 734static struct gfs2_dirent *gfs2_init_dirent(struct inode *inode,
 735					    struct gfs2_dirent *dent,
 736					    const struct qstr *name,
 737					    struct buffer_head *bh)
 738{
 739	unsigned offset = 0;
 740
 741	if (!gfs2_dirent_sentinel(dent))
 742		offset = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
 743	return do_init_dirent(inode, dent, name, bh, offset);
 744}
 745
 746static struct gfs2_dirent *gfs2_dirent_split_alloc(struct inode *inode,
 747						   struct buffer_head *bh,
 748						   const struct qstr *name,
 749						   void *ptr)
 750{
 751	struct gfs2_dirent *dent;
 752	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
 753				gfs2_dirent_find_offset, name, ptr);
 754	if (!dent || IS_ERR(dent))
 755		return dent;
 756	return do_init_dirent(inode, dent, name, bh,
 757			      (unsigned)(ptr - (void *)dent));
 758}
 759
 760static int get_leaf(struct gfs2_inode *dip, u64 leaf_no,
 761		    struct buffer_head **bhp)
 762{
 763	int error;
 764
 765	error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, 0, bhp);
 766	if (!error && gfs2_metatype_check(GFS2_SB(&dip->i_inode), *bhp, GFS2_METATYPE_LF)) {
 767		/* pr_info("block num=%llu\n", leaf_no); */
 768		error = -EIO;
 769	}
 770
 771	return error;
 772}
 773
 774/**
 775 * get_leaf_nr - Get a leaf number associated with the index
 776 * @dip: The GFS2 inode
 777 * @index:
 778 * @leaf_out:
 779 *
 780 * Returns: 0 on success, error code otherwise
 781 */
 782
 783static int get_leaf_nr(struct gfs2_inode *dip, u32 index,
 784		       u64 *leaf_out)
 785{
 786	__be64 *hash;
 787	int error;
 788
 789	hash = gfs2_dir_get_hash_table(dip);
 790	error = PTR_ERR_OR_ZERO(hash);
 791
 792	if (!error)
 793		*leaf_out = be64_to_cpu(*(hash + index));
 794
 795	return error;
 796}
 797
 798static int get_first_leaf(struct gfs2_inode *dip, u32 index,
 799			  struct buffer_head **bh_out)
 800{
 801	u64 leaf_no;
 802	int error;
 803
 804	error = get_leaf_nr(dip, index, &leaf_no);
 805	if (!error)
 806		error = get_leaf(dip, leaf_no, bh_out);
 807
 808	return error;
 809}
 810
 811static struct gfs2_dirent *gfs2_dirent_search(struct inode *inode,
 812					      const struct qstr *name,
 813					      gfs2_dscan_t scan,
 814					      struct buffer_head **pbh)
 815{
 816	struct buffer_head *bh;
 817	struct gfs2_dirent *dent;
 818	struct gfs2_inode *ip = GFS2_I(inode);
 819	int error;
 820
 821	if (ip->i_diskflags & GFS2_DIF_EXHASH) {
 822		struct gfs2_leaf *leaf;
 823		unsigned int hsize = BIT(ip->i_depth);
 824		unsigned int index;
 825		u64 ln;
 826		if (hsize * sizeof(u64) != i_size_read(inode)) {
 827			gfs2_consist_inode(ip);
 828			return ERR_PTR(-EIO);
 829		}
 830
 831		index = name->hash >> (32 - ip->i_depth);
 832		error = get_first_leaf(ip, index, &bh);
 833		if (error)
 834			return ERR_PTR(error);
 835		do {
 836			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
 837						scan, name, NULL);
 838			if (dent)
 839				goto got_dent;
 840			leaf = (struct gfs2_leaf *)bh->b_data;
 841			ln = be64_to_cpu(leaf->lf_next);
 842			brelse(bh);
 843			if (!ln)
 844				break;
 845
 846			error = get_leaf(ip, ln, &bh);
 847		} while(!error);
 848
 849		return error ? ERR_PTR(error) : NULL;
 850	}
 851
 852
 853	error = gfs2_meta_inode_buffer(ip, &bh);
 854	if (error)
 855		return ERR_PTR(error);
 856	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL);
 857got_dent:
 858	if (unlikely(dent == NULL || IS_ERR(dent))) {
 859		brelse(bh);
 860		bh = NULL;
 861	}
 862	*pbh = bh;
 863	return dent;
 864}
 865
 866static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth)
 867{
 868	struct gfs2_inode *ip = GFS2_I(inode);
 869	unsigned int n = 1;
 870	u64 bn;
 871	int error;
 872	struct buffer_head *bh;
 873	struct gfs2_leaf *leaf;
 874	struct gfs2_dirent *dent;
 875	struct timespec tv = current_time(inode);
 876
 877	error = gfs2_alloc_blocks(ip, &bn, &n, 0, NULL);
 878	if (error)
 879		return NULL;
 880	bh = gfs2_meta_new(ip->i_gl, bn);
 881	if (!bh)
 882		return NULL;
 883
 884	gfs2_trans_add_unrevoke(GFS2_SB(inode), bn, 1);
 885	gfs2_trans_add_meta(ip->i_gl, bh);
 886	gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF);
 887	leaf = (struct gfs2_leaf *)bh->b_data;
 888	leaf->lf_depth = cpu_to_be16(depth);
 889	leaf->lf_entries = 0;
 890	leaf->lf_dirent_format = cpu_to_be32(GFS2_FORMAT_DE);
 891	leaf->lf_next = 0;
 892	leaf->lf_inode = cpu_to_be64(ip->i_no_addr);
 893	leaf->lf_dist = cpu_to_be32(1);
 894	leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
 895	leaf->lf_sec = cpu_to_be64(tv.tv_sec);
 896	memset(leaf->lf_reserved2, 0, sizeof(leaf->lf_reserved2));
 897	dent = (struct gfs2_dirent *)(leaf+1);
 898	gfs2_qstr2dirent(&empty_name, bh->b_size - sizeof(struct gfs2_leaf), dent);
 899	*pbh = bh;
 900	return leaf;
 901}
 902
 903/**
 904 * dir_make_exhash - Convert a stuffed directory into an ExHash directory
 905 * @dip: The GFS2 inode
 906 *
 907 * Returns: 0 on success, error code otherwise
 908 */
 909
 910static int dir_make_exhash(struct inode *inode)
 911{
 912	struct gfs2_inode *dip = GFS2_I(inode);
 913	struct gfs2_sbd *sdp = GFS2_SB(inode);
 914	struct gfs2_dirent *dent;
 915	struct qstr args;
 916	struct buffer_head *bh, *dibh;
 917	struct gfs2_leaf *leaf;
 918	int y;
 919	u32 x;
 920	__be64 *lp;
 921	u64 bn;
 922	int error;
 923
 924	error = gfs2_meta_inode_buffer(dip, &dibh);
 925	if (error)
 926		return error;
 927
 928	/*  Turn over a new leaf  */
 929
 930	leaf = new_leaf(inode, &bh, 0);
 931	if (!leaf)
 932		return -ENOSPC;
 933	bn = bh->b_blocknr;
 934
 935	gfs2_assert(sdp, dip->i_entries < BIT(16));
 936	leaf->lf_entries = cpu_to_be16(dip->i_entries);
 937
 938	/*  Copy dirents  */
 939
 940	gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh,
 941			     sizeof(struct gfs2_dinode));
 942
 943	/*  Find last entry  */
 944
 945	x = 0;
 946	args.len = bh->b_size - sizeof(struct gfs2_dinode) +
 947		   sizeof(struct gfs2_leaf);
 948	args.name = bh->b_data;
 949	dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size,
 950				gfs2_dirent_last, &args, NULL);
 951	if (!dent) {
 952		brelse(bh);
 953		brelse(dibh);
 954		return -EIO;
 955	}
 956	if (IS_ERR(dent)) {
 957		brelse(bh);
 958		brelse(dibh);
 959		return PTR_ERR(dent);
 960	}
 961
 962	/*  Adjust the last dirent's record length
 963	   (Remember that dent still points to the last entry.)  */
 964
 965	dent->de_rec_len = cpu_to_be16(be16_to_cpu(dent->de_rec_len) +
 966		sizeof(struct gfs2_dinode) -
 967		sizeof(struct gfs2_leaf));
 968
 969	brelse(bh);
 970
 971	/*  We're done with the new leaf block, now setup the new
 972	    hash table.  */
 973
 974	gfs2_trans_add_meta(dip->i_gl, dibh);
 975	gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode));
 976
 977	lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode));
 978
 979	for (x = sdp->sd_hash_ptrs; x--; lp++)
 980		*lp = cpu_to_be64(bn);
 981
 982	i_size_write(inode, sdp->sd_sb.sb_bsize / 2);
 983	gfs2_add_inode_blocks(&dip->i_inode, 1);
 984	dip->i_diskflags |= GFS2_DIF_EXHASH;
 985
 986	for (x = sdp->sd_hash_ptrs, y = -1; x; x >>= 1, y++) ;
 987	dip->i_depth = y;
 988
 989	gfs2_dinode_out(dip, dibh->b_data);
 990
 991	brelse(dibh);
 992
 993	return 0;
 994}
 995
 996/**
 997 * dir_split_leaf - Split a leaf block into two
 998 * @dip: The GFS2 inode
 999 * @index:
1000 * @leaf_no:
1001 *
1002 * Returns: 0 on success, error code on failure
1003 */
1004
1005static int dir_split_leaf(struct inode *inode, const struct qstr *name)
1006{
1007	struct gfs2_inode *dip = GFS2_I(inode);
1008	struct buffer_head *nbh, *obh, *dibh;
1009	struct gfs2_leaf *nleaf, *oleaf;
1010	struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
1011	u32 start, len, half_len, divider;
1012	u64 bn, leaf_no;
1013	__be64 *lp;
1014	u32 index;
1015	int x, moved = 0;
1016	int error;
1017
1018	index = name->hash >> (32 - dip->i_depth);
1019	error = get_leaf_nr(dip, index, &leaf_no);
1020	if (error)
1021		return error;
1022
1023	/*  Get the old leaf block  */
1024	error = get_leaf(dip, leaf_no, &obh);
1025	if (error)
1026		return error;
1027
1028	oleaf = (struct gfs2_leaf *)obh->b_data;
1029	if (dip->i_depth == be16_to_cpu(oleaf->lf_depth)) {
1030		brelse(obh);
1031		return 1; /* can't split */
1032	}
1033
1034	gfs2_trans_add_meta(dip->i_gl, obh);
1035
1036	nleaf = new_leaf(inode, &nbh, be16_to_cpu(oleaf->lf_depth) + 1);
1037	if (!nleaf) {
1038		brelse(obh);
1039		return -ENOSPC;
1040	}
1041	bn = nbh->b_blocknr;
1042
1043	/*  Compute the start and len of leaf pointers in the hash table.  */
1044	len = BIT(dip->i_depth - be16_to_cpu(oleaf->lf_depth));
1045	half_len = len >> 1;
1046	if (!half_len) {
1047		pr_warn("i_depth %u lf_depth %u index %u\n",
1048			dip->i_depth, be16_to_cpu(oleaf->lf_depth), index);
1049		gfs2_consist_inode(dip);
1050		error = -EIO;
1051		goto fail_brelse;
1052	}
1053
1054	start = (index & ~(len - 1));
1055
1056	/* Change the pointers.
1057	   Don't bother distinguishing stuffed from non-stuffed.
1058	   This code is complicated enough already. */
1059	lp = kmalloc(half_len * sizeof(__be64), GFP_NOFS);
1060	if (!lp) {
1061		error = -ENOMEM;
1062		goto fail_brelse;
1063	}
1064
1065	/*  Change the pointers  */
1066	for (x = 0; x < half_len; x++)
1067		lp[x] = cpu_to_be64(bn);
1068
1069	gfs2_dir_hash_inval(dip);
1070
1071	error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64),
1072				    half_len * sizeof(u64));
1073	if (error != half_len * sizeof(u64)) {
1074		if (error >= 0)
1075			error = -EIO;
1076		goto fail_lpfree;
1077	}
1078
1079	kfree(lp);
1080
1081	/*  Compute the divider  */
1082	divider = (start + half_len) << (32 - dip->i_depth);
1083
1084	/*  Copy the entries  */
1085	dent = (struct gfs2_dirent *)(obh->b_data + sizeof(struct gfs2_leaf));
1086
1087	do {
1088		next = dent;
1089		if (dirent_next(dip, obh, &next))
1090			next = NULL;
1091
1092		if (!gfs2_dirent_sentinel(dent) &&
1093		    be32_to_cpu(dent->de_hash) < divider) {
1094			struct qstr str;
1095			void *ptr = ((char *)dent - obh->b_data) + nbh->b_data;
1096			str.name = (char*)(dent+1);
1097			str.len = be16_to_cpu(dent->de_name_len);
1098			str.hash = be32_to_cpu(dent->de_hash);
1099			new = gfs2_dirent_split_alloc(inode, nbh, &str, ptr);
1100			if (IS_ERR(new)) {
1101				error = PTR_ERR(new);
1102				break;
1103			}
1104
1105			new->de_inum = dent->de_inum; /* No endian worries */
1106			new->de_type = dent->de_type; /* No endian worries */
1107			be16_add_cpu(&nleaf->lf_entries, 1);
1108
1109			dirent_del(dip, obh, prev, dent);
1110
1111			if (!oleaf->lf_entries)
1112				gfs2_consist_inode(dip);
1113			be16_add_cpu(&oleaf->lf_entries, -1);
1114
1115			if (!prev)
1116				prev = dent;
1117
1118			moved = 1;
1119		} else {
1120			prev = dent;
1121		}
1122		dent = next;
1123	} while (dent);
1124
1125	oleaf->lf_depth = nleaf->lf_depth;
1126
1127	error = gfs2_meta_inode_buffer(dip, &dibh);
1128	if (!gfs2_assert_withdraw(GFS2_SB(&dip->i_inode), !error)) {
1129		gfs2_trans_add_meta(dip->i_gl, dibh);
1130		gfs2_add_inode_blocks(&dip->i_inode, 1);
1131		gfs2_dinode_out(dip, dibh->b_data);
1132		brelse(dibh);
1133	}
1134
1135	brelse(obh);
1136	brelse(nbh);
1137
1138	return error;
1139
1140fail_lpfree:
1141	kfree(lp);
1142
1143fail_brelse:
1144	brelse(obh);
1145	brelse(nbh);
1146	return error;
1147}
1148
1149/**
1150 * dir_double_exhash - Double size of ExHash table
1151 * @dip: The GFS2 dinode
1152 *
1153 * Returns: 0 on success, error code on failure
1154 */
1155
1156static int dir_double_exhash(struct gfs2_inode *dip)
1157{
1158	struct buffer_head *dibh;
1159	u32 hsize;
1160	u32 hsize_bytes;
1161	__be64 *hc;
1162	__be64 *hc2, *h;
1163	int x;
1164	int error = 0;
1165
1166	hsize = BIT(dip->i_depth);
1167	hsize_bytes = hsize * sizeof(__be64);
1168
1169	hc = gfs2_dir_get_hash_table(dip);
1170	if (IS_ERR(hc))
1171		return PTR_ERR(hc);
1172
1173	hc2 = kmalloc(hsize_bytes * 2, GFP_NOFS | __GFP_NOWARN);
1174	if (hc2 == NULL)
1175		hc2 = __vmalloc(hsize_bytes * 2, GFP_NOFS, PAGE_KERNEL);
1176
1177	if (!hc2)
1178		return -ENOMEM;
1179
1180	h = hc2;
1181	error = gfs2_meta_inode_buffer(dip, &dibh);
1182	if (error)
1183		goto out_kfree;
1184
1185	for (x = 0; x < hsize; x++) {
1186		*h++ = *hc;
1187		*h++ = *hc;
1188		hc++;
1189	}
1190
1191	error = gfs2_dir_write_data(dip, (char *)hc2, 0, hsize_bytes * 2);
1192	if (error != (hsize_bytes * 2))
1193		goto fail;
1194
1195	gfs2_dir_hash_inval(dip);
1196	dip->i_hash_cache = hc2;
1197	dip->i_depth++;
1198	gfs2_dinode_out(dip, dibh->b_data);
1199	brelse(dibh);
1200	return 0;
1201
1202fail:
1203	/* Replace original hash table & size */
1204	gfs2_dir_write_data(dip, (char *)hc, 0, hsize_bytes);
1205	i_size_write(&dip->i_inode, hsize_bytes);
1206	gfs2_dinode_out(dip, dibh->b_data);
1207	brelse(dibh);
1208out_kfree:
1209	kvfree(hc2);
1210	return error;
1211}
1212
1213/**
1214 * compare_dents - compare directory entries by hash value
1215 * @a: first dent
1216 * @b: second dent
1217 *
1218 * When comparing the hash entries of @a to @b:
1219 *   gt: returns 1
1220 *   lt: returns -1
1221 *   eq: returns 0
1222 */
1223
1224static int compare_dents(const void *a, const void *b)
1225{
1226	const struct gfs2_dirent *dent_a, *dent_b;
1227	u32 hash_a, hash_b;
1228	int ret = 0;
1229
1230	dent_a = *(const struct gfs2_dirent **)a;
1231	hash_a = dent_a->de_cookie;
1232
1233	dent_b = *(const struct gfs2_dirent **)b;
1234	hash_b = dent_b->de_cookie;
1235
1236	if (hash_a > hash_b)
1237		ret = 1;
1238	else if (hash_a < hash_b)
1239		ret = -1;
1240	else {
1241		unsigned int len_a = be16_to_cpu(dent_a->de_name_len);
1242		unsigned int len_b = be16_to_cpu(dent_b->de_name_len);
1243
1244		if (len_a > len_b)
1245			ret = 1;
1246		else if (len_a < len_b)
1247			ret = -1;
1248		else
1249			ret = memcmp(dent_a + 1, dent_b + 1, len_a);
1250	}
1251
1252	return ret;
1253}
1254
1255/**
1256 * do_filldir_main - read out directory entries
1257 * @dip: The GFS2 inode
1258 * @ctx: what to feed the entries to
1259 * @darr: an array of struct gfs2_dirent pointers to read
1260 * @entries: the number of entries in darr
1261 * @copied: pointer to int that's non-zero if a entry has been copied out
1262 *
1263 * Jump through some hoops to make sure that if there are hash collsions,
1264 * they are read out at the beginning of a buffer.  We want to minimize
1265 * the possibility that they will fall into different readdir buffers or
1266 * that someone will want to seek to that location.
1267 *
1268 * Returns: errno, >0 if the actor tells you to stop
1269 */
1270
1271static int do_filldir_main(struct gfs2_inode *dip, struct dir_context *ctx,
1272			   struct gfs2_dirent **darr, u32 entries,
1273			   u32 sort_start, int *copied)
1274{
1275	const struct gfs2_dirent *dent, *dent_next;
1276	u64 off, off_next;
1277	unsigned int x, y;
1278	int run = 0;
1279
1280	if (sort_start < entries)
1281		sort(&darr[sort_start], entries - sort_start,
1282		     sizeof(struct gfs2_dirent *), compare_dents, NULL);
1283
1284	dent_next = darr[0];
1285	off_next = dent_next->de_cookie;
1286
1287	for (x = 0, y = 1; x < entries; x++, y++) {
1288		dent = dent_next;
1289		off = off_next;
1290
1291		if (y < entries) {
1292			dent_next = darr[y];
1293			off_next = dent_next->de_cookie;
1294
1295			if (off < ctx->pos)
1296				continue;
1297			ctx->pos = off;
1298
1299			if (off_next == off) {
1300				if (*copied && !run)
1301					return 1;
1302				run = 1;
1303			} else
1304				run = 0;
1305		} else {
1306			if (off < ctx->pos)
1307				continue;
1308			ctx->pos = off;
1309		}
1310
1311		if (!dir_emit(ctx, (const char *)(dent + 1),
1312				be16_to_cpu(dent->de_name_len),
1313				be64_to_cpu(dent->de_inum.no_addr),
1314				be16_to_cpu(dent->de_type)))
1315			return 1;
1316
1317		*copied = 1;
1318	}
1319
1320	/* Increment the ctx->pos by one, so the next time we come into the
1321	   do_filldir fxn, we get the next entry instead of the last one in the
1322	   current leaf */
1323
1324	ctx->pos++;
1325
1326	return 0;
1327}
1328
1329static void *gfs2_alloc_sort_buffer(unsigned size)
1330{
1331	void *ptr = NULL;
1332
1333	if (size < KMALLOC_MAX_SIZE)
1334		ptr = kmalloc(size, GFP_NOFS | __GFP_NOWARN);
1335	if (!ptr)
1336		ptr = __vmalloc(size, GFP_NOFS, PAGE_KERNEL);
1337	return ptr;
1338}
1339
1340
1341static int gfs2_set_cookies(struct gfs2_sbd *sdp, struct buffer_head *bh,
1342			    unsigned leaf_nr, struct gfs2_dirent **darr,
1343			    unsigned entries)
1344{
1345	int sort_id = -1;
1346	int i;
1347	
1348	for (i = 0; i < entries; i++) {
1349		unsigned offset;
1350
1351		darr[i]->de_cookie = be32_to_cpu(darr[i]->de_hash);
1352		darr[i]->de_cookie = gfs2_disk_hash2offset(darr[i]->de_cookie);
1353
1354		if (!sdp->sd_args.ar_loccookie)
1355			continue;
1356		offset = (char *)(darr[i]) -
1357			 (bh->b_data + gfs2_dirent_offset(bh->b_data));
1358		offset /= GFS2_MIN_DIRENT_SIZE;
1359		offset += leaf_nr * sdp->sd_max_dents_per_leaf;
1360		if (offset >= GFS2_USE_HASH_FLAG ||
1361		    leaf_nr >= GFS2_USE_HASH_FLAG) {
1362			darr[i]->de_cookie |= GFS2_USE_HASH_FLAG;
1363			if (sort_id < 0)
1364				sort_id = i;
1365			continue;
1366		}
1367		darr[i]->de_cookie &= GFS2_HASH_INDEX_MASK;
1368		darr[i]->de_cookie |= offset;
1369	}
1370	return sort_id;
1371}	
1372
1373
1374static int gfs2_dir_read_leaf(struct inode *inode, struct dir_context *ctx,
1375			      int *copied, unsigned *depth,
1376			      u64 leaf_no)
1377{
1378	struct gfs2_inode *ip = GFS2_I(inode);
1379	struct gfs2_sbd *sdp = GFS2_SB(inode);
1380	struct buffer_head *bh;
1381	struct gfs2_leaf *lf;
1382	unsigned entries = 0, entries2 = 0;
1383	unsigned leaves = 0, leaf = 0, offset, sort_offset;
1384	struct gfs2_dirent **darr, *dent;
1385	struct dirent_gather g;
1386	struct buffer_head **larr;
1387	int error, i, need_sort = 0, sort_id;
1388	u64 lfn = leaf_no;
1389
1390	do {
1391		error = get_leaf(ip, lfn, &bh);
1392		if (error)
1393			goto out;
1394		lf = (struct gfs2_leaf *)bh->b_data;
1395		if (leaves == 0)
1396			*depth = be16_to_cpu(lf->lf_depth);
1397		entries += be16_to_cpu(lf->lf_entries);
1398		leaves++;
1399		lfn = be64_to_cpu(lf->lf_next);
1400		brelse(bh);
1401	} while(lfn);
1402
1403	if (*depth < GFS2_DIR_MAX_DEPTH || !sdp->sd_args.ar_loccookie) {
1404		need_sort = 1;
1405		sort_offset = 0;
1406	}
1407
1408	if (!entries)
1409		return 0;
1410
1411	error = -ENOMEM;
1412	/*
1413	 * The extra 99 entries are not normally used, but are a buffer
1414	 * zone in case the number of entries in the leaf is corrupt.
1415	 * 99 is the maximum number of entries that can fit in a single
1416	 * leaf block.
1417	 */
1418	larr = gfs2_alloc_sort_buffer((leaves + entries + 99) * sizeof(void *));
1419	if (!larr)
1420		goto out;
1421	darr = (struct gfs2_dirent **)(larr + leaves);
1422	g.pdent = (const struct gfs2_dirent **)darr;
1423	g.offset = 0;
1424	lfn = leaf_no;
1425
1426	do {
1427		error = get_leaf(ip, lfn, &bh);
1428		if (error)
1429			goto out_free;
1430		lf = (struct gfs2_leaf *)bh->b_data;
1431		lfn = be64_to_cpu(lf->lf_next);
1432		if (lf->lf_entries) {
1433			offset = g.offset;
1434			entries2 += be16_to_cpu(lf->lf_entries);
1435			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
1436						gfs2_dirent_gather, NULL, &g);
1437			error = PTR_ERR(dent);
1438			if (IS_ERR(dent))
1439				goto out_free;
1440			if (entries2 != g.offset) {
1441				fs_warn(sdp, "Number of entries corrupt in dir "
1442						"leaf %llu, entries2 (%u) != "
1443						"g.offset (%u)\n",
1444					(unsigned long long)bh->b_blocknr,
1445					entries2, g.offset);
1446				gfs2_consist_inode(ip);
1447				error = -EIO;
1448				goto out_free;
1449			}
1450			error = 0;
1451			sort_id = gfs2_set_cookies(sdp, bh, leaf, &darr[offset],
1452						   be16_to_cpu(lf->lf_entries));
1453			if (!need_sort && sort_id >= 0) {
1454				need_sort = 1;
1455				sort_offset = offset + sort_id;
1456			}
1457			larr[leaf++] = bh;
1458		} else {
1459			larr[leaf++] = NULL;
1460			brelse(bh);
1461		}
1462	} while(lfn);
1463
1464	BUG_ON(entries2 != entries);
1465	error = do_filldir_main(ip, ctx, darr, entries, need_sort ?
1466				sort_offset : entries, copied);
1467out_free:
1468	for(i = 0; i < leaf; i++)
1469		if (larr[i])
1470			brelse(larr[i]);
1471	kvfree(larr);
1472out:
1473	return error;
1474}
1475
1476/**
1477 * gfs2_dir_readahead - Issue read-ahead requests for leaf blocks.
1478 *
1479 * Note: we can't calculate each index like dir_e_read can because we don't
1480 * have the leaf, and therefore we don't have the depth, and therefore we
1481 * don't have the length. So we have to just read enough ahead to make up
1482 * for the loss of information.
1483 */
1484static void gfs2_dir_readahead(struct inode *inode, unsigned hsize, u32 index,
1485			       struct file_ra_state *f_ra)
1486{
1487	struct gfs2_inode *ip = GFS2_I(inode);
1488	struct gfs2_glock *gl = ip->i_gl;
1489	struct buffer_head *bh;
1490	u64 blocknr = 0, last;
1491	unsigned count;
1492
1493	/* First check if we've already read-ahead for the whole range. */
1494	if (index + MAX_RA_BLOCKS < f_ra->start)
1495		return;
1496
1497	f_ra->start = max((pgoff_t)index, f_ra->start);
1498	for (count = 0; count < MAX_RA_BLOCKS; count++) {
1499		if (f_ra->start >= hsize) /* if exceeded the hash table */
1500			break;
1501
1502		last = blocknr;
1503		blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]);
1504		f_ra->start++;
1505		if (blocknr == last)
1506			continue;
1507
1508		bh = gfs2_getbuf(gl, blocknr, 1);
1509		if (trylock_buffer(bh)) {
1510			if (buffer_uptodate(bh)) {
1511				unlock_buffer(bh);
1512				brelse(bh);
1513				continue;
1514			}
1515			bh->b_end_io = end_buffer_read_sync;
1516			submit_bh(REQ_OP_READ,
1517				  REQ_RAHEAD | REQ_META | REQ_PRIO,
1518				  bh);
1519			continue;
1520		}
1521		brelse(bh);
1522	}
1523}
1524
1525/**
1526 * dir_e_read - Reads the entries from a directory into a filldir buffer
1527 * @dip: dinode pointer
1528 * @ctx: actor to feed the entries to
1529 *
1530 * Returns: errno
1531 */
1532
1533static int dir_e_read(struct inode *inode, struct dir_context *ctx,
1534		      struct file_ra_state *f_ra)
1535{
1536	struct gfs2_inode *dip = GFS2_I(inode);
1537	u32 hsize, len = 0;
1538	u32 hash, index;
1539	__be64 *lp;
1540	int copied = 0;
1541	int error = 0;
1542	unsigned depth = 0;
1543
1544	hsize = BIT(dip->i_depth);
1545	hash = gfs2_dir_offset2hash(ctx->pos);
1546	index = hash >> (32 - dip->i_depth);
1547
1548	if (dip->i_hash_cache == NULL)
1549		f_ra->start = 0;
1550	lp = gfs2_dir_get_hash_table(dip);
1551	if (IS_ERR(lp))
1552		return PTR_ERR(lp);
1553
1554	gfs2_dir_readahead(inode, hsize, index, f_ra);
1555
1556	while (index < hsize) {
1557		error = gfs2_dir_read_leaf(inode, ctx,
1558					   &copied, &depth,
1559					   be64_to_cpu(lp[index]));
1560		if (error)
1561			break;
1562
1563		len = BIT(dip->i_depth - depth);
1564		index = (index & ~(len - 1)) + len;
1565	}
1566
1567	if (error > 0)
1568		error = 0;
1569	return error;
1570}
1571
1572int gfs2_dir_read(struct inode *inode, struct dir_context *ctx,
1573		  struct file_ra_state *f_ra)
1574{
1575	struct gfs2_inode *dip = GFS2_I(inode);
1576	struct gfs2_sbd *sdp = GFS2_SB(inode);
1577	struct dirent_gather g;
1578	struct gfs2_dirent **darr, *dent;
1579	struct buffer_head *dibh;
1580	int copied = 0;
1581	int error;
1582
1583	if (!dip->i_entries)
1584		return 0;
1585
1586	if (dip->i_diskflags & GFS2_DIF_EXHASH)
1587		return dir_e_read(inode, ctx, f_ra);
1588
1589	if (!gfs2_is_stuffed(dip)) {
1590		gfs2_consist_inode(dip);
1591		return -EIO;
1592	}
1593
1594	error = gfs2_meta_inode_buffer(dip, &dibh);
1595	if (error)
1596		return error;
1597
1598	error = -ENOMEM;
1599	/* 96 is max number of dirents which can be stuffed into an inode */
1600	darr = kmalloc(96 * sizeof(struct gfs2_dirent *), GFP_NOFS);
1601	if (darr) {
1602		g.pdent = (const struct gfs2_dirent **)darr;
1603		g.offset = 0;
1604		dent = gfs2_dirent_scan(inode, dibh->b_data, dibh->b_size,
1605					gfs2_dirent_gather, NULL, &g);
1606		if (IS_ERR(dent)) {
1607			error = PTR_ERR(dent);
1608			goto out;
1609		}
1610		if (dip->i_entries != g.offset) {
1611			fs_warn(sdp, "Number of entries corrupt in dir %llu, "
1612				"ip->i_entries (%u) != g.offset (%u)\n",
1613				(unsigned long long)dip->i_no_addr,
1614				dip->i_entries,
1615				g.offset);
1616			gfs2_consist_inode(dip);
1617			error = -EIO;
1618			goto out;
1619		}
1620		gfs2_set_cookies(sdp, dibh, 0, darr, dip->i_entries);
1621		error = do_filldir_main(dip, ctx, darr,
1622					dip->i_entries, 0, &copied);
1623out:
1624		kfree(darr);
1625	}
1626
1627	if (error > 0)
1628		error = 0;
1629
1630	brelse(dibh);
1631
1632	return error;
1633}
1634
1635/**
1636 * gfs2_dir_search - Search a directory
1637 * @dip: The GFS2 dir inode
1638 * @name: The name we are looking up
1639 * @fail_on_exist: Fail if the name exists rather than looking it up
1640 *
1641 * This routine searches a directory for a file or another directory.
1642 * Assumes a glock is held on dip.
1643 *
1644 * Returns: errno
1645 */
1646
1647struct inode *gfs2_dir_search(struct inode *dir, const struct qstr *name,
1648			      bool fail_on_exist)
1649{
1650	struct buffer_head *bh;
1651	struct gfs2_dirent *dent;
1652	u64 addr, formal_ino;
1653	u16 dtype;
1654
1655	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1656	if (dent) {
1657		struct inode *inode;
1658		u16 rahead;
1659
1660		if (IS_ERR(dent))
1661			return ERR_CAST(dent);
1662		dtype = be16_to_cpu(dent->de_type);
1663		rahead = be16_to_cpu(dent->de_rahead);
1664		addr = be64_to_cpu(dent->de_inum.no_addr);
1665		formal_ino = be64_to_cpu(dent->de_inum.no_formal_ino);
1666		brelse(bh);
1667		if (fail_on_exist)
1668			return ERR_PTR(-EEXIST);
1669		inode = gfs2_inode_lookup(dir->i_sb, dtype, addr, formal_ino,
1670					  GFS2_BLKST_FREE /* ignore */);
1671		if (!IS_ERR(inode))
1672			GFS2_I(inode)->i_rahead = rahead;
1673		return inode;
1674	}
1675	return ERR_PTR(-ENOENT);
1676}
1677
1678int gfs2_dir_check(struct inode *dir, const struct qstr *name,
1679		   const struct gfs2_inode *ip)
1680{
1681	struct buffer_head *bh;
1682	struct gfs2_dirent *dent;
1683	int ret = -ENOENT;
1684
1685	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1686	if (dent) {
1687		if (IS_ERR(dent))
1688			return PTR_ERR(dent);
1689		if (ip) {
1690			if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr)
1691				goto out;
1692			if (be64_to_cpu(dent->de_inum.no_formal_ino) !=
1693			    ip->i_no_formal_ino)
1694				goto out;
1695			if (unlikely(IF2DT(ip->i_inode.i_mode) !=
1696			    be16_to_cpu(dent->de_type))) {
1697				gfs2_consist_inode(GFS2_I(dir));
1698				ret = -EIO;
1699				goto out;
1700			}
1701		}
1702		ret = 0;
1703out:
1704		brelse(bh);
1705	}
1706	return ret;
1707}
1708
1709/**
1710 * dir_new_leaf - Add a new leaf onto hash chain
1711 * @inode: The directory
1712 * @name: The name we are adding
1713 *
1714 * This adds a new dir leaf onto an existing leaf when there is not
1715 * enough space to add a new dir entry. This is a last resort after
1716 * we've expanded the hash table to max size and also split existing
1717 * leaf blocks, so it will only occur for very large directories.
1718 *
1719 * The dist parameter is set to 1 for leaf blocks directly attached
1720 * to the hash table, 2 for one layer of indirection, 3 for two layers
1721 * etc. We are thus able to tell the difference between an old leaf
1722 * with dist set to zero (i.e. "don't know") and a new one where we
1723 * set this information for debug/fsck purposes.
1724 *
1725 * Returns: 0 on success, or -ve on error
1726 */
1727
1728static int dir_new_leaf(struct inode *inode, const struct qstr *name)
1729{
1730	struct buffer_head *bh, *obh;
1731	struct gfs2_inode *ip = GFS2_I(inode);
1732	struct gfs2_leaf *leaf, *oleaf;
1733	u32 dist = 1;
1734	int error;
1735	u32 index;
1736	u64 bn;
1737
1738	index = name->hash >> (32 - ip->i_depth);
1739	error = get_first_leaf(ip, index, &obh);
1740	if (error)
1741		return error;
1742	do {
1743		dist++;
1744		oleaf = (struct gfs2_leaf *)obh->b_data;
1745		bn = be64_to_cpu(oleaf->lf_next);
1746		if (!bn)
1747			break;
1748		brelse(obh);
1749		error = get_leaf(ip, bn, &obh);
1750		if (error)
1751			return error;
1752	} while(1);
1753
1754	gfs2_trans_add_meta(ip->i_gl, obh);
1755
1756	leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth));
1757	if (!leaf) {
1758		brelse(obh);
1759		return -ENOSPC;
1760	}
1761	leaf->lf_dist = cpu_to_be32(dist);
1762	oleaf->lf_next = cpu_to_be64(bh->b_blocknr);
1763	brelse(bh);
1764	brelse(obh);
1765
1766	error = gfs2_meta_inode_buffer(ip, &bh);
1767	if (error)
1768		return error;
1769	gfs2_trans_add_meta(ip->i_gl, bh);
1770	gfs2_add_inode_blocks(&ip->i_inode, 1);
1771	gfs2_dinode_out(ip, bh->b_data);
1772	brelse(bh);
1773	return 0;
1774}
1775
1776static u16 gfs2_inode_ra_len(const struct gfs2_inode *ip)
1777{
1778	u64 where = ip->i_no_addr + 1;
1779	if (ip->i_eattr == where)
1780		return 1;
1781	return 0;
1782}
1783
1784/**
1785 * gfs2_dir_add - Add new filename into directory
1786 * @inode: The directory inode
1787 * @name: The new name
1788 * @nip: The GFS2 inode to be linked in to the directory
1789 * @da: The directory addition info
1790 *
1791 * If the call to gfs2_diradd_alloc_required resulted in there being
1792 * no need to allocate any new directory blocks, then it will contain
1793 * a pointer to the directory entry and the bh in which it resides. We
1794 * can use that without having to repeat the search. If there was no
1795 * free space, then we must now create more space.
1796 *
1797 * Returns: 0 on success, error code on failure
1798 */
1799
1800int gfs2_dir_add(struct inode *inode, const struct qstr *name,
1801		 const struct gfs2_inode *nip, struct gfs2_diradd *da)
1802{
1803	struct gfs2_inode *ip = GFS2_I(inode);
1804	struct buffer_head *bh = da->bh;
1805	struct gfs2_dirent *dent = da->dent;
1806	struct timespec tv;
1807	struct gfs2_leaf *leaf;
1808	int error;
1809
1810	while(1) {
1811		if (da->bh == NULL) {
1812			dent = gfs2_dirent_search(inode, name,
1813						  gfs2_dirent_find_space, &bh);
1814		}
1815		if (dent) {
1816			if (IS_ERR(dent))
1817				return PTR_ERR(dent);
1818			dent = gfs2_init_dirent(inode, dent, name, bh);
1819			gfs2_inum_out(nip, dent);
1820			dent->de_type = cpu_to_be16(IF2DT(nip->i_inode.i_mode));
1821			dent->de_rahead = cpu_to_be16(gfs2_inode_ra_len(nip));
1822			tv = current_time(&ip->i_inode);
1823			if (ip->i_diskflags & GFS2_DIF_EXHASH) {
1824				leaf = (struct gfs2_leaf *)bh->b_data;
1825				be16_add_cpu(&leaf->lf_entries, 1);
1826				leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1827				leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1828			}
1829			da->dent = NULL;
1830			da->bh = NULL;
1831			brelse(bh);
1832			ip->i_entries++;
1833			ip->i_inode.i_mtime = ip->i_inode.i_ctime = tv;
1834			if (S_ISDIR(nip->i_inode.i_mode))
1835				inc_nlink(&ip->i_inode);
1836			mark_inode_dirty(inode);
1837			error = 0;
1838			break;
1839		}
1840		if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) {
1841			error = dir_make_exhash(inode);
1842			if (error)
1843				break;
1844			continue;
1845		}
1846		error = dir_split_leaf(inode, name);
1847		if (error == 0)
1848			continue;
1849		if (error < 0)
1850			break;
1851		if (ip->i_depth < GFS2_DIR_MAX_DEPTH) {
1852			error = dir_double_exhash(ip);
1853			if (error)
1854				break;
1855			error = dir_split_leaf(inode, name);
1856			if (error < 0)
1857				break;
1858			if (error == 0)
1859				continue;
1860		}
1861		error = dir_new_leaf(inode, name);
1862		if (!error)
1863			continue;
1864		error = -ENOSPC;
1865		break;
1866	}
1867	return error;
1868}
1869
1870
1871/**
1872 * gfs2_dir_del - Delete a directory entry
1873 * @dip: The GFS2 inode
1874 * @filename: The filename
1875 *
1876 * Returns: 0 on success, error code on failure
1877 */
1878
1879int gfs2_dir_del(struct gfs2_inode *dip, const struct dentry *dentry)
1880{
1881	const struct qstr *name = &dentry->d_name;
1882	struct gfs2_dirent *dent, *prev = NULL;
1883	struct buffer_head *bh;
1884	struct timespec tv = current_time(&dip->i_inode);
1885
1886	/* Returns _either_ the entry (if its first in block) or the
1887	   previous entry otherwise */
1888	dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh);
1889	if (!dent) {
1890		gfs2_consist_inode(dip);
1891		return -EIO;
1892	}
1893	if (IS_ERR(dent)) {
1894		gfs2_consist_inode(dip);
1895		return PTR_ERR(dent);
1896	}
1897	/* If not first in block, adjust pointers accordingly */
1898	if (gfs2_dirent_find(dent, name, NULL) == 0) {
1899		prev = dent;
1900		dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len));
1901	}
1902
1903	dirent_del(dip, bh, prev, dent);
1904	if (dip->i_diskflags & GFS2_DIF_EXHASH) {
1905		struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data;
1906		u16 entries = be16_to_cpu(leaf->lf_entries);
1907		if (!entries)
1908			gfs2_consist_inode(dip);
1909		leaf->lf_entries = cpu_to_be16(--entries);
1910		leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1911		leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1912	}
1913	brelse(bh);
1914
1915	if (!dip->i_entries)
1916		gfs2_consist_inode(dip);
1917	dip->i_entries--;
1918	dip->i_inode.i_mtime = dip->i_inode.i_ctime = tv;
1919	if (d_is_dir(dentry))
1920		drop_nlink(&dip->i_inode);
1921	mark_inode_dirty(&dip->i_inode);
1922
1923	return 0;
1924}
1925
1926/**
1927 * gfs2_dir_mvino - Change inode number of directory entry
1928 * @dip: The GFS2 inode
1929 * @filename:
1930 * @new_inode:
1931 *
1932 * This routine changes the inode number of a directory entry.  It's used
1933 * by rename to change ".." when a directory is moved.
1934 * Assumes a glock is held on dvp.
1935 *
1936 * Returns: errno
1937 */
1938
1939int gfs2_dir_mvino(struct gfs2_inode *dip, const struct qstr *filename,
1940		   const struct gfs2_inode *nip, unsigned int new_type)
1941{
1942	struct buffer_head *bh;
1943	struct gfs2_dirent *dent;
1944	int error;
1945
1946	dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh);
1947	if (!dent) {
1948		gfs2_consist_inode(dip);
1949		return -EIO;
1950	}
1951	if (IS_ERR(dent))
1952		return PTR_ERR(dent);
1953
1954	gfs2_trans_add_meta(dip->i_gl, bh);
1955	gfs2_inum_out(nip, dent);
1956	dent->de_type = cpu_to_be16(new_type);
1957
1958	if (dip->i_diskflags & GFS2_DIF_EXHASH) {
1959		brelse(bh);
1960		error = gfs2_meta_inode_buffer(dip, &bh);
1961		if (error)
1962			return error;
1963		gfs2_trans_add_meta(dip->i_gl, bh);
1964	}
1965
1966	dip->i_inode.i_mtime = dip->i_inode.i_ctime = current_time(&dip->i_inode);
1967	gfs2_dinode_out(dip, bh->b_data);
1968	brelse(bh);
1969	return 0;
1970}
1971
1972/**
1973 * leaf_dealloc - Deallocate a directory leaf
1974 * @dip: the directory
1975 * @index: the hash table offset in the directory
1976 * @len: the number of pointers to this leaf
1977 * @leaf_no: the leaf number
1978 * @leaf_bh: buffer_head for the starting leaf
1979 * last_dealloc: 1 if this is the final dealloc for the leaf, else 0
1980 *
1981 * Returns: errno
1982 */
1983
1984static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len,
1985			u64 leaf_no, struct buffer_head *leaf_bh,
1986			int last_dealloc)
1987{
1988	struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode);
1989	struct gfs2_leaf *tmp_leaf;
1990	struct gfs2_rgrp_list rlist;
1991	struct buffer_head *bh, *dibh;
1992	u64 blk, nblk;
1993	unsigned int rg_blocks = 0, l_blocks = 0;
1994	char *ht;
1995	unsigned int x, size = len * sizeof(u64);
1996	int error;
1997
1998	error = gfs2_rindex_update(sdp);
1999	if (error)
2000		return error;
2001
2002	memset(&rlist, 0, sizeof(struct gfs2_rgrp_list));
2003
2004	ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN);
2005	if (ht == NULL)
2006		ht = __vmalloc(size, GFP_NOFS | __GFP_NOWARN | __GFP_ZERO,
2007			       PAGE_KERNEL);
2008	if (!ht)
2009		return -ENOMEM;
2010
2011	error = gfs2_quota_hold(dip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
2012	if (error)
2013		goto out;
2014
2015	/*  Count the number of leaves  */
2016	bh = leaf_bh;
2017
2018	for (blk = leaf_no; blk; blk = nblk) {
2019		if (blk != leaf_no) {
2020			error = get_leaf(dip, blk, &bh);
2021			if (error)
2022				goto out_rlist;
2023		}
2024		tmp_leaf = (struct gfs2_leaf *)bh->b_data;
2025		nblk = be64_to_cpu(tmp_leaf->lf_next);
2026		if (blk != leaf_no)
2027			brelse(bh);
2028
2029		gfs2_rlist_add(dip, &rlist, blk);
2030		l_blocks++;
2031	}
2032
2033	gfs2_rlist_alloc(&rlist, LM_ST_EXCLUSIVE);
2034
2035	for (x = 0; x < rlist.rl_rgrps; x++) {
2036		struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(rlist.rl_ghs[x].gh_gl);
2037
2038		rg_blocks += rgd->rd_length;
2039	}
2040
2041	error = gfs2_glo

Large files files are truncated, but you can click here to view the full file