|
@@ -503,7 +503,6 @@ int fat_alloc_clusters(struct inode *inode, int *cluster, int nr_cluster)
|
|
|
sbi->prev_free = entry;
|
|
|
if (sbi->free_clusters != -1)
|
|
|
sbi->free_clusters--;
|
|
|
- mark_fsinfo_dirty(sb);
|
|
|
|
|
|
cluster[idx_clus] = entry;
|
|
|
idx_clus++;
|
|
@@ -525,11 +524,11 @@ int fat_alloc_clusters(struct inode *inode, int *cluster, int nr_cluster)
|
|
|
/* Couldn't allocate the free entries */
|
|
|
sbi->free_clusters = 0;
|
|
|
sbi->free_clus_valid = 1;
|
|
|
- mark_fsinfo_dirty(sb);
|
|
|
err = -ENOSPC;
|
|
|
|
|
|
out:
|
|
|
unlock_fat(sbi);
|
|
|
+ mark_fsinfo_dirty(sb);
|
|
|
fatent_brelse(&fatent);
|
|
|
if (!err) {
|
|
|
if (inode_needs_sync(inode))
|
|
@@ -554,7 +553,7 @@ int fat_free_clusters(struct inode *inode, int cluster)
|
|
|
struct fat_entry fatent;
|
|
|
struct buffer_head *bhs[MAX_BUF_PER_PAGE];
|
|
|
int i, err, nr_bhs;
|
|
|
- int first_cl = cluster;
|
|
|
+ int first_cl = cluster, dirty_fsinfo = 0;
|
|
|
|
|
|
nr_bhs = 0;
|
|
|
fatent_init(&fatent);
|
|
@@ -592,7 +591,7 @@ int fat_free_clusters(struct inode *inode, int cluster)
|
|
|
ops->ent_put(&fatent, FAT_ENT_FREE);
|
|
|
if (sbi->free_clusters != -1) {
|
|
|
sbi->free_clusters++;
|
|
|
- mark_fsinfo_dirty(sb);
|
|
|
+ dirty_fsinfo = 1;
|
|
|
}
|
|
|
|
|
|
if (nr_bhs + fatent.nr_bhs > MAX_BUF_PER_PAGE) {
|
|
@@ -622,6 +621,8 @@ error:
|
|
|
for (i = 0; i < nr_bhs; i++)
|
|
|
brelse(bhs[i]);
|
|
|
unlock_fat(sbi);
|
|
|
+ if (dirty_fsinfo)
|
|
|
+ mark_fsinfo_dirty(sb);
|
|
|
|
|
|
return err;
|
|
|
}
|