2007-06-12 07:07:21 -06:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2007 Oracle. All rights reserved.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public
|
|
|
|
* License v2 as published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public
|
|
|
|
* License along with this program; if not, write to the
|
|
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
|
|
* Boston, MA 021110-1307, USA.
|
|
|
|
*/
|
|
|
|
|
2008-01-08 13:46:30 -07:00
|
|
|
#ifndef __BTRFS_TRANSACTION__
|
|
|
|
#define __BTRFS_TRANSACTION__
|
2007-04-30 13:25:45 -06:00
|
|
|
#include "btrfs_inode.h"
|
2009-03-13 08:10:06 -06:00
|
|
|
#include "delayed-ref.h"
|
2007-03-16 14:20:31 -06:00
|
|
|
|
2007-03-22 13:59:16 -06:00
|
|
|
struct btrfs_transaction {
|
|
|
|
u64 transid;
|
2009-03-12 18:12:45 -06:00
|
|
|
/*
|
|
|
|
* total writers in this transaction, it must be zero before the
|
|
|
|
* transaction can end
|
|
|
|
*/
|
2011-04-11 13:45:29 -06:00
|
|
|
atomic_t num_writers;
|
2009-03-12 18:12:45 -06:00
|
|
|
|
2007-08-10 14:22:09 -06:00
|
|
|
unsigned long num_joined;
|
2007-03-22 13:59:16 -06:00
|
|
|
int in_commit;
|
2011-04-11 13:45:29 -06:00
|
|
|
atomic_t use_count;
|
2007-03-22 13:59:16 -06:00
|
|
|
int commit_done;
|
2008-07-17 10:54:14 -06:00
|
|
|
int blocked;
|
2007-04-19 19:01:03 -06:00
|
|
|
struct list_head list;
|
2008-01-24 14:13:08 -07:00
|
|
|
struct extent_io_tree dirty_pages;
|
2007-06-08 13:33:54 -06:00
|
|
|
unsigned long start_time;
|
2007-03-22 13:59:16 -06:00
|
|
|
wait_queue_head_t writer_wait;
|
|
|
|
wait_queue_head_t commit_wait;
|
2008-01-08 13:46:30 -07:00
|
|
|
struct list_head pending_snapshots;
|
2009-03-13 08:10:06 -06:00
|
|
|
struct btrfs_delayed_ref_root delayed_refs;
|
2007-03-22 13:59:16 -06:00
|
|
|
};
|
|
|
|
|
2007-03-16 14:20:31 -06:00
|
|
|
struct btrfs_trans_handle {
|
|
|
|
u64 transid;
|
2010-05-16 08:46:25 -06:00
|
|
|
u64 block_group;
|
|
|
|
u64 bytes_reserved;
|
2007-03-16 14:20:31 -06:00
|
|
|
unsigned long blocks_reserved;
|
|
|
|
unsigned long blocks_used;
|
2009-03-13 08:10:06 -06:00
|
|
|
unsigned long delayed_ref_updates;
|
2010-05-16 08:46:25 -06:00
|
|
|
struct btrfs_transaction *transaction;
|
|
|
|
struct btrfs_block_rsv *block_rsv;
|
2007-03-16 14:20:31 -06:00
|
|
|
};
|
|
|
|
|
2008-01-08 13:46:30 -07:00
|
|
|
struct btrfs_pending_snapshot {
|
2008-11-17 19:02:50 -07:00
|
|
|
struct dentry *dentry;
|
2008-01-08 13:46:30 -07:00
|
|
|
struct btrfs_root *root;
|
2010-05-16 08:48:46 -06:00
|
|
|
struct btrfs_root *snap;
|
|
|
|
/* block reservation for the operation */
|
|
|
|
struct btrfs_block_rsv block_rsv;
|
|
|
|
/* extra metadata reseration for relocation */
|
|
|
|
int error;
|
2010-12-20 01:04:08 -07:00
|
|
|
bool readonly;
|
2008-01-08 13:46:30 -07:00
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
2007-04-30 13:25:45 -06:00
|
|
|
static inline void btrfs_set_trans_block_group(struct btrfs_trans_handle *trans,
|
|
|
|
struct inode *inode)
|
|
|
|
{
|
|
|
|
trans->block_group = BTRFS_I(inode)->block_group;
|
|
|
|
}
|
|
|
|
|
2009-01-05 19:25:51 -07:00
|
|
|
static inline void btrfs_update_inode_block_group(
|
|
|
|
struct btrfs_trans_handle *trans,
|
|
|
|
struct inode *inode)
|
2007-04-30 13:25:45 -06:00
|
|
|
{
|
|
|
|
BTRFS_I(inode)->block_group = trans->block_group;
|
|
|
|
}
|
|
|
|
|
2007-08-10 14:22:09 -06:00
|
|
|
static inline void btrfs_set_inode_last_trans(struct btrfs_trans_handle *trans,
|
|
|
|
struct inode *inode)
|
|
|
|
{
|
|
|
|
BTRFS_I(inode)->last_trans = trans->transaction->transid;
|
2009-10-13 11:21:08 -06:00
|
|
|
BTRFS_I(inode)->last_sub_trans = BTRFS_I(inode)->root->log_transid;
|
2007-08-10 14:22:09 -06:00
|
|
|
}
|
|
|
|
|
2007-03-22 13:59:16 -06:00
|
|
|
int btrfs_end_transaction(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2010-06-21 12:48:16 -06:00
|
|
|
int btrfs_end_transaction_nolock(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2007-03-22 13:59:16 -06:00
|
|
|
struct btrfs_trans_handle *btrfs_start_transaction(struct btrfs_root *root,
|
2010-05-16 08:48:46 -06:00
|
|
|
int num_items);
|
2008-07-17 10:54:14 -06:00
|
|
|
struct btrfs_trans_handle *btrfs_join_transaction(struct btrfs_root *root,
|
2010-05-16 08:48:46 -06:00
|
|
|
int num_blocks);
|
2010-06-21 12:48:16 -06:00
|
|
|
struct btrfs_trans_handle *btrfs_join_transaction_nolock(struct btrfs_root *root,
|
|
|
|
int num_blocks);
|
2008-08-04 08:41:27 -06:00
|
|
|
struct btrfs_trans_handle *btrfs_start_ioctl_transaction(struct btrfs_root *r,
|
2010-05-16 08:48:46 -06:00
|
|
|
int num_blocks);
|
Btrfs: add START_SYNC, WAIT_SYNC ioctls
START_SYNC will start a sync/commit, but not wait for it to
complete. Any modification started after the ioctl returns is
guaranteed not to be included in the commit. If a non-NULL
pointer is passed, the transaction id will be returned to
userspace.
WAIT_SYNC will wait for any in-progress commit to complete. If a
transaction id is specified, the ioctl will block and then
return (success) when the specified transaction has committed.
If it has already committed when we call the ioctl, it returns
immediately. If the specified transaction doesn't exist, it
returns EINVAL.
If no transaction id is specified, WAIT_SYNC will wait for the
currently committing transaction to finish it's commit to disk.
If there is no currently committing transaction, it returns
success.
These ioctls are useful for applications which want to impose an
ordering on when fs modifications reach disk, but do not want to
wait for the full (slow) commit process to do so.
Picky callers can take the transid returned by START_SYNC and
feed it to WAIT_SYNC, and be certain to wait only as long as
necessary for the transaction _they_ started to reach disk.
Sloppy callers can START_SYNC and WAIT_SYNC without a transid,
and provided they didn't wait too long between the calls, they
will get the same result. However, if a second commit starts
before they call WAIT_SYNC, they may end up waiting longer for
it to commit as well. Even so, a START_SYNC+WAIT_SYNC still
guarantees that any operation completed before the START_SYNC
reaches disk.
Signed-off-by: Sage Weil <sage@newdream.net>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2010-10-29 13:41:32 -06:00
|
|
|
int btrfs_wait_for_commit(struct btrfs_root *root, u64 transid);
|
2007-03-22 13:59:16 -06:00
|
|
|
int btrfs_write_and_wait_transaction(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
|
|
|
int btrfs_commit_tree_roots(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2007-06-08 13:33:54 -06:00
|
|
|
|
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 08:45:14 -06:00
|
|
|
int btrfs_add_dead_root(struct btrfs_root *root);
|
|
|
|
int btrfs_drop_dead_root(struct btrfs_root *root);
|
2007-08-10 12:06:19 -06:00
|
|
|
int btrfs_defrag_root(struct btrfs_root *root, int cacheonly);
|
|
|
|
int btrfs_clean_old_snapshots(struct btrfs_root *root);
|
2007-10-15 14:14:19 -06:00
|
|
|
int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2010-10-29 13:37:34 -06:00
|
|
|
int btrfs_commit_transaction_async(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root,
|
|
|
|
int wait_for_unblock);
|
2008-06-25 14:01:31 -06:00
|
|
|
int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
btrfs: implement delayed inode items operation
Changelog V5 -> V6:
- Fix oom when the memory load is high, by storing the delayed nodes into the
root's radix tree, and letting btrfs inodes go.
Changelog V4 -> V5:
- Fix the race on adding the delayed node to the inode, which is spotted by
Chris Mason.
- Merge Chris Mason's incremental patch into this patch.
- Fix deadlock between readdir() and memory fault, which is reported by
Itaru Kitayama.
Changelog V3 -> V4:
- Fix nested lock, which is reported by Itaru Kitayama, by updating space cache
inode in time.
Changelog V2 -> V3:
- Fix the race between the delayed worker and the task which does delayed items
balance, which is reported by Tsutomu Itoh.
- Modify the patch address David Sterba's comment.
- Fix the bug of the cpu recursion spinlock, reported by Chris Mason
Changelog V1 -> V2:
- break up the global rb-tree, use a list to manage the delayed nodes,
which is created for every directory and file, and used to manage the
delayed directory name index items and the delayed inode item.
- introduce a worker to deal with the delayed nodes.
Compare with Ext3/4, the performance of file creation and deletion on btrfs
is very poor. the reason is that btrfs must do a lot of b+ tree insertions,
such as inode item, directory name item, directory name index and so on.
If we can do some delayed b+ tree insertion or deletion, we can improve the
performance, so we made this patch which implemented delayed directory name
index insertion/deletion and delayed inode update.
Implementation:
- introduce a delayed root object into the filesystem, that use two lists to
manage the delayed nodes which are created for every file/directory.
One is used to manage all the delayed nodes that have delayed items. And the
other is used to manage the delayed nodes which is waiting to be dealt with
by the work thread.
- Every delayed node has two rb-tree, one is used to manage the directory name
index which is going to be inserted into b+ tree, and the other is used to
manage the directory name index which is going to be deleted from b+ tree.
- introduce a worker to deal with the delayed operation. This worker is used
to deal with the works of the delayed directory name index items insertion
and deletion and the delayed inode update.
When the delayed items is beyond the lower limit, we create works for some
delayed nodes and insert them into the work queue of the worker, and then
go back.
When the delayed items is beyond the upper bound, we create works for all
the delayed nodes that haven't been dealt with, and insert them into the work
queue of the worker, and then wait for that the untreated items is below some
threshold value.
- When we want to insert a directory name index into b+ tree, we just add the
information into the delayed inserting rb-tree.
And then we check the number of the delayed items and do delayed items
balance. (The balance policy is above.)
- When we want to delete a directory name index from the b+ tree, we search it
in the inserting rb-tree at first. If we look it up, just drop it. If not,
add the key of it into the delayed deleting rb-tree.
Similar to the delayed inserting rb-tree, we also check the number of the
delayed items and do delayed items balance.
(The same to inserting manipulation)
- When we want to update the metadata of some inode, we cached the data of the
inode into the delayed node. the worker will flush it into the b+ tree after
dealing with the delayed insertion and deletion.
- We will move the delayed node to the tail of the list after we access the
delayed node, By this way, we can cache more delayed items and merge more
inode updates.
- If we want to commit transaction, we will deal with all the delayed node.
- the delayed node will be freed when we free the btrfs inode.
- Before we log the inode items, we commit all the directory name index items
and the delayed inode update.
I did a quick test by the benchmark tool[1] and found we can improve the
performance of file creation by ~15%, and file deletion by ~20%.
Before applying this patch:
Create files:
Total files: 50000
Total time: 1.096108
Average time: 0.000022
Delete files:
Total files: 50000
Total time: 1.510403
Average time: 0.000030
After applying this patch:
Create files:
Total files: 50000
Total time: 0.932899
Average time: 0.000019
Delete files:
Total files: 50000
Total time: 1.215732
Average time: 0.000024
[1] http://marc.info/?l=linux-btrfs&m=128212635122920&q=p3
Many thanks for Kitayama-san's help!
Signed-off-by: Miao Xie <miaox@cn.fujitsu.com>
Reviewed-by: David Sterba <dave@jikos.cz>
Tested-by: Tsutomu Itoh <t-itoh@jp.fujitsu.com>
Tested-by: Itaru Kitayama <kitayama@cl.bb4u.ne.jp>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2011-04-22 04:12:22 -06:00
|
|
|
int btrfs_end_transaction_dmeta(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2010-05-16 08:49:58 -06:00
|
|
|
int btrfs_should_end_transaction(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2008-07-29 14:15:18 -06:00
|
|
|
void btrfs_throttle(struct btrfs_root *root);
|
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 08:45:14 -06:00
|
|
|
int btrfs_record_root_in_trans(struct btrfs_trans_handle *trans,
|
|
|
|
struct btrfs_root *root);
|
2008-09-11 14:17:57 -06:00
|
|
|
int btrfs_write_and_wait_marked_extents(struct btrfs_root *root,
|
2009-11-12 02:33:26 -07:00
|
|
|
struct extent_io_tree *dirty_pages, int mark);
|
2009-10-13 11:29:19 -06:00
|
|
|
int btrfs_write_marked_extents(struct btrfs_root *root,
|
2009-11-12 02:33:26 -07:00
|
|
|
struct extent_io_tree *dirty_pages, int mark);
|
2009-10-13 11:29:19 -06:00
|
|
|
int btrfs_wait_marked_extents(struct btrfs_root *root,
|
2009-11-12 02:33:26 -07:00
|
|
|
struct extent_io_tree *dirty_pages, int mark);
|
2010-05-16 08:49:58 -06:00
|
|
|
int btrfs_transaction_blocked(struct btrfs_fs_info *info);
|
2009-07-30 08:04:48 -06:00
|
|
|
int btrfs_transaction_in_commit(struct btrfs_fs_info *info);
|
2007-03-16 14:20:31 -06:00
|
|
|
#endif
|