1
0
Fork 0
mirror of https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git synced 2025-01-22 16:06:04 -05:00

ufs cleanups, fixes and folio conversion

Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
 -----BEGIN PGP SIGNATURE-----
 
 iHUEABYIAB0WIQQqUNBr3gm4hGXdBJlZ7Krx/gZQ6wUCZzdkpwAKCRBZ7Krx/gZQ
 632xAP0WukRCsROhDX1LsNXY+SF3E6znsO8VwjjHTDwZ+dGW6gD/f+KooaGG5qBU
 4Z7sVuOIpOOA0S1LLDx8XBYidcG23g8=
 =Aw1m
 -----END PGP SIGNATURE-----

Merge tag 'pull-ufs' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs

Pull ufs updates from Al Viro:
 "ufs cleanups, fixes and folio conversion"

* tag 'pull-ufs' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs:
  ufs: ufs_sb_private_info: remove unused s_{2,3}apb fields
  ufs: Convert ufs_change_blocknr() to take a folio
  ufs: Pass a folio to ufs_new_fragments()
  ufs: Convert ufs_inode_getfrag() to take a folio
  ufs: Convert ufs_extend_tail() to take a folio
  ufs: Convert ufs_inode_getblock() to take a folio
  ufs: take the handling of free block counters into a helper
  clean ufs_trunc_direct() up a bit...
  ufs: get rid of ubh_{ubhcpymem,memcpyubh}()
  ufs_inode_getfrag(): remove junk comment
  ufs_free_fragments(): fix the braino in sanity check
  ufs_clusteracct(): switch to passing fragment number
  ufs: untangle ubh_...block...(), part 3
  ufs: untangle ubh_...block...(), part 2
  ufs: untangle ubh_...block...() macros, part 1
  ufs: fix ufs_read_cylinder() failure handling
  ufs: missing ->splice_write()
  ufs: fix handling of delete_entry and set_link failures
This commit is contained in:
Linus Torvalds 2024-11-18 12:58:23 -08:00
commit 9fb2cfa463
11 changed files with 225 additions and 339 deletions

View file

@ -33,6 +33,29 @@ static u64 ufs_bitmap_search (struct super_block *, struct ufs_cg_private_info *
static unsigned char ufs_fragtable_8fpb[], ufs_fragtable_other[]; static unsigned char ufs_fragtable_8fpb[], ufs_fragtable_other[];
static void ufs_clusteracct(struct super_block *, struct ufs_cg_private_info *, unsigned, int); static void ufs_clusteracct(struct super_block *, struct ufs_cg_private_info *, unsigned, int);
static void adjust_free_blocks(struct super_block *sb,
struct ufs_cylinder_group *ucg,
struct ufs_cg_private_info *ucpi,
unsigned fragment, int delta)
{
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
ufs_clusteracct(sb, ucpi, fragment, delta);
fs32_add(sb, &ucg->cg_cs.cs_nbfree, delta);
uspi->cs_total.cs_nbfree += delta;
fs32_add(sb, &UFS_SB(sb)->fs_cs(ucpi->c_cgx).cs_nbfree, delta);
if (uspi->fs_magic != UFS2_MAGIC) {
unsigned cylno = ufs_cbtocylno(fragment);
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
ufs_cbtorpos(fragment)), delta);
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), delta);
}
}
/* /*
* Free 'count' fragments from fragment number 'fragment' * Free 'count' fragments from fragment number 'fragment'
*/ */
@ -43,7 +66,6 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
struct ufs_cg_private_info * ucpi; struct ufs_cg_private_info * ucpi;
struct ufs_cylinder_group * ucg; struct ufs_cylinder_group * ucg;
unsigned cgno, bit, end_bit, bbase, blkmap, i; unsigned cgno, bit, end_bit, bbase, blkmap, i;
u64 blkno;
sb = inode->i_sb; sb = inode->i_sb;
uspi = UFS_SB(sb)->s_uspi; uspi = UFS_SB(sb)->s_uspi;
@ -51,7 +73,7 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
UFSD("ENTER, fragment %llu, count %u\n", UFSD("ENTER, fragment %llu, count %u\n",
(unsigned long long)fragment, count); (unsigned long long)fragment, count);
if (ufs_fragnum(fragment) + count > uspi->s_fpg) if (ufs_fragnum(fragment) + count > uspi->s_fpb)
ufs_error (sb, "ufs_free_fragments", "internal error"); ufs_error (sb, "ufs_free_fragments", "internal error");
mutex_lock(&UFS_SB(sb)->s_lock); mutex_lock(&UFS_SB(sb)->s_lock);
@ -94,23 +116,11 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
/* /*
* Trying to reassemble free fragments into block * Trying to reassemble free fragments into block
*/ */
blkno = ufs_fragstoblks (bbase); if (ubh_isblockset(uspi, ucpi, bbase)) {
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno)) {
fs32_sub(sb, &ucg->cg_cs.cs_nffree, uspi->s_fpb); fs32_sub(sb, &ucg->cg_cs.cs_nffree, uspi->s_fpb);
uspi->cs_total.cs_nffree -= uspi->s_fpb; uspi->cs_total.cs_nffree -= uspi->s_fpb;
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, uspi->s_fpb); fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, uspi->s_fpb);
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD) adjust_free_blocks(sb, ucg, ucpi, bbase, 1);
ufs_clusteracct (sb, ucpi, blkno, 1);
fs32_add(sb, &ucg->cg_cs.cs_nbfree, 1);
uspi->cs_total.cs_nbfree++;
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nbfree, 1);
if (uspi->fs_magic != UFS2_MAGIC) {
unsigned cylno = ufs_cbtocylno (bbase);
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
ufs_cbtorpos(bbase)), 1);
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), 1);
}
} }
ubh_mark_buffer_dirty (USPI_UBH(uspi)); ubh_mark_buffer_dirty (USPI_UBH(uspi));
@ -139,7 +149,6 @@ void ufs_free_blocks(struct inode *inode, u64 fragment, unsigned count)
struct ufs_cg_private_info * ucpi; struct ufs_cg_private_info * ucpi;
struct ufs_cylinder_group * ucg; struct ufs_cylinder_group * ucg;
unsigned overflow, cgno, bit, end_bit, i; unsigned overflow, cgno, bit, end_bit, i;
u64 blkno;
sb = inode->i_sb; sb = inode->i_sb;
uspi = UFS_SB(sb)->s_uspi; uspi = UFS_SB(sb)->s_uspi;
@ -181,26 +190,12 @@ void ufs_free_blocks(struct inode *inode, u64 fragment, unsigned count)
} }
for (i = bit; i < end_bit; i += uspi->s_fpb) { for (i = bit; i < end_bit; i += uspi->s_fpb) {
blkno = ufs_fragstoblks(i); if (ubh_isblockset(uspi, ucpi, i)) {
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno)) {
ufs_error(sb, "ufs_free_blocks", "freeing free fragment"); ufs_error(sb, "ufs_free_blocks", "freeing free fragment");
} }
ubh_setblock(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno); ubh_setblock(uspi, ucpi, i);
inode_sub_bytes(inode, uspi->s_fpb << uspi->s_fshift); inode_sub_bytes(inode, uspi->s_fpb << uspi->s_fshift);
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD) adjust_free_blocks(sb, ucg, ucpi, i, 1);
ufs_clusteracct (sb, ucpi, blkno, 1);
fs32_add(sb, &ucg->cg_cs.cs_nbfree, 1);
uspi->cs_total.cs_nbfree++;
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nbfree, 1);
if (uspi->fs_magic != UFS2_MAGIC) {
unsigned cylno = ufs_cbtocylno(i);
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
ufs_cbtorpos(i)), 1);
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), 1);
}
} }
ubh_mark_buffer_dirty (USPI_UBH(uspi)); ubh_mark_buffer_dirty (USPI_UBH(uspi));
@ -234,13 +229,13 @@ void ufs_free_blocks(struct inode *inode, u64 fragment, unsigned count)
* situated at the end of file. * situated at the end of file.
* *
* We can come here from ufs_writepage or ufs_prepare_write, * We can come here from ufs_writepage or ufs_prepare_write,
* locked_page is argument of these functions, so we already lock it. * locked_folio is argument of these functions, so we already lock it.
*/ */
static void ufs_change_blocknr(struct inode *inode, sector_t beg, static void ufs_change_blocknr(struct inode *inode, sector_t beg,
unsigned int count, sector_t oldb, unsigned int count, sector_t oldb,
sector_t newb, struct page *locked_page) sector_t newb, struct folio *locked_folio)
{ {
struct folio *folio, *locked_folio = page_folio(locked_page); struct folio *folio;
const unsigned blks_per_page = const unsigned blks_per_page =
1 << (PAGE_SHIFT - inode->i_blkbits); 1 << (PAGE_SHIFT - inode->i_blkbits);
const unsigned mask = blks_per_page - 1; const unsigned mask = blks_per_page - 1;
@ -337,7 +332,7 @@ static void ufs_clear_frags(struct inode *inode, sector_t beg, unsigned int n,
u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment, u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
u64 goal, unsigned count, int *err, u64 goal, unsigned count, int *err,
struct page *locked_page) struct folio *locked_folio)
{ {
struct super_block * sb; struct super_block * sb;
struct ufs_sb_private_info * uspi; struct ufs_sb_private_info * uspi;
@ -417,7 +412,7 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
result = ufs_alloc_fragments (inode, cgno, goal, count, err); result = ufs_alloc_fragments (inode, cgno, goal, count, err);
if (result) { if (result) {
ufs_clear_frags(inode, result + oldcount, ufs_clear_frags(inode, result + oldcount,
newcount - oldcount, locked_page != NULL); newcount - oldcount, locked_folio != NULL);
*err = 0; *err = 0;
write_seqlock(&UFS_I(inode)->meta_lock); write_seqlock(&UFS_I(inode)->meta_lock);
ufs_cpu_to_data_ptr(sb, p, result); ufs_cpu_to_data_ptr(sb, p, result);
@ -441,7 +436,7 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
fragment + count); fragment + count);
read_sequnlock_excl(&UFS_I(inode)->meta_lock); read_sequnlock_excl(&UFS_I(inode)->meta_lock);
ufs_clear_frags(inode, result + oldcount, newcount - oldcount, ufs_clear_frags(inode, result + oldcount, newcount - oldcount,
locked_page != NULL); locked_folio != NULL);
mutex_unlock(&UFS_SB(sb)->s_lock); mutex_unlock(&UFS_SB(sb)->s_lock);
UFSD("EXIT, result %llu\n", (unsigned long long)result); UFSD("EXIT, result %llu\n", (unsigned long long)result);
return result; return result;
@ -462,11 +457,11 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
result = ufs_alloc_fragments (inode, cgno, goal, request, err); result = ufs_alloc_fragments (inode, cgno, goal, request, err);
if (result) { if (result) {
ufs_clear_frags(inode, result + oldcount, newcount - oldcount, ufs_clear_frags(inode, result + oldcount, newcount - oldcount,
locked_page != NULL); locked_folio != NULL);
mutex_unlock(&UFS_SB(sb)->s_lock); mutex_unlock(&UFS_SB(sb)->s_lock);
ufs_change_blocknr(inode, fragment - oldcount, oldcount, ufs_change_blocknr(inode, fragment - oldcount, oldcount,
uspi->s_sbbase + tmp, uspi->s_sbbase + tmp,
uspi->s_sbbase + result, locked_page); uspi->s_sbbase + result, locked_folio);
*err = 0; *err = 0;
write_seqlock(&UFS_I(inode)->meta_lock); write_seqlock(&UFS_I(inode)->meta_lock);
ufs_cpu_to_data_ptr(sb, p, result); ufs_cpu_to_data_ptr(sb, p, result);
@ -698,7 +693,7 @@ static u64 ufs_alloccg_block(struct inode *inode,
struct super_block * sb; struct super_block * sb;
struct ufs_sb_private_info * uspi; struct ufs_sb_private_info * uspi;
struct ufs_cylinder_group * ucg; struct ufs_cylinder_group * ucg;
u64 result, blkno; u64 result;
UFSD("ENTER, goal %llu\n", (unsigned long long)goal); UFSD("ENTER, goal %llu\n", (unsigned long long)goal);
@ -716,7 +711,7 @@ static u64 ufs_alloccg_block(struct inode *inode,
/* /*
* If the requested block is available, use it. * If the requested block is available, use it.
*/ */
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, ufs_fragstoblks(goal))) { if (ubh_isblockset(uspi, ucpi, goal)) {
result = goal; result = goal;
goto gotit; goto gotit;
} }
@ -729,22 +724,8 @@ static u64 ufs_alloccg_block(struct inode *inode,
gotit: gotit:
if (!try_add_frags(inode, uspi->s_fpb)) if (!try_add_frags(inode, uspi->s_fpb))
return 0; return 0;
blkno = ufs_fragstoblks(result); ubh_clrblock(uspi, ucpi, result);
ubh_clrblock (UCPI_UBH(ucpi), ucpi->c_freeoff, blkno); adjust_free_blocks(sb, ucg, ucpi, result, -1);
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
ufs_clusteracct (sb, ucpi, blkno, -1);
fs32_sub(sb, &ucg->cg_cs.cs_nbfree, 1);
uspi->cs_total.cs_nbfree--;
fs32_sub(sb, &UFS_SB(sb)->fs_cs(ucpi->c_cgx).cs_nbfree, 1);
if (uspi->fs_magic != UFS2_MAGIC) {
unsigned cylno = ufs_cbtocylno((unsigned)result);
fs16_sub(sb, &ubh_cg_blks(ucpi, cylno,
ufs_cbtorpos((unsigned)result)), 1);
fs32_sub(sb, &ubh_cg_blktot(ucpi, cylno), 1);
}
UFSD("EXIT, result %llu\n", (unsigned long long)result); UFSD("EXIT, result %llu\n", (unsigned long long)result);
@ -863,12 +844,12 @@ static u64 ufs_bitmap_search(struct super_block *sb,
} }
static void ufs_clusteracct(struct super_block * sb, static void ufs_clusteracct(struct super_block * sb,
struct ufs_cg_private_info * ucpi, unsigned blkno, int cnt) struct ufs_cg_private_info * ucpi, unsigned frag, int cnt)
{ {
struct ufs_sb_private_info * uspi; struct ufs_sb_private_info * uspi = UFS_SB(sb)->s_uspi;
int i, start, end, forw, back; int i, start, end, forw, back;
unsigned blkno = ufs_fragstoblks(frag);
uspi = UFS_SB(sb)->s_uspi;
if (uspi->s_contigsumsize <= 0) if (uspi->s_contigsumsize <= 0)
return; return;

View file

@ -26,7 +26,7 @@
* Read cylinder group into cache. The memory space for ufs_cg_private_info * Read cylinder group into cache. The memory space for ufs_cg_private_info
* structure is already allocated during ufs_read_super. * structure is already allocated during ufs_read_super.
*/ */
static void ufs_read_cylinder (struct super_block * sb, static bool ufs_read_cylinder(struct super_block *sb,
unsigned cgno, unsigned bitmap_nr) unsigned cgno, unsigned bitmap_nr)
{ {
struct ufs_sb_info * sbi = UFS_SB(sb); struct ufs_sb_info * sbi = UFS_SB(sb);
@ -46,9 +46,11 @@ static void ufs_read_cylinder (struct super_block * sb,
* We have already the first fragment of cylinder group block in buffer * We have already the first fragment of cylinder group block in buffer
*/ */
UCPI_UBH(ucpi)->bh[0] = sbi->s_ucg[cgno]; UCPI_UBH(ucpi)->bh[0] = sbi->s_ucg[cgno];
for (i = 1; i < UCPI_UBH(ucpi)->count; i++) for (i = 1; i < UCPI_UBH(ucpi)->count; i++) {
if (!(UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i))) UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i);
if (!UCPI_UBH(ucpi)->bh[i])
goto failed; goto failed;
}
sbi->s_cgno[bitmap_nr] = cgno; sbi->s_cgno[bitmap_nr] = cgno;
ucpi->c_cgx = fs32_to_cpu(sb, ucg->cg_cgx); ucpi->c_cgx = fs32_to_cpu(sb, ucg->cg_cgx);
@ -67,13 +69,14 @@ static void ufs_read_cylinder (struct super_block * sb,
ucpi->c_clusteroff = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_clusteroff); ucpi->c_clusteroff = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_clusteroff);
ucpi->c_nclusterblks = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_nclusterblks); ucpi->c_nclusterblks = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_nclusterblks);
UFSD("EXIT\n"); UFSD("EXIT\n");
return; return true;
failed: failed:
for (j = 1; j < i; j++) for (j = 1; j < i; j++)
brelse (sbi->s_ucg[j]); brelse(UCPI_UBH(ucpi)->bh[j]);
sbi->s_cgno[bitmap_nr] = UFS_CGNO_EMPTY; sbi->s_cgno[bitmap_nr] = UFS_CGNO_EMPTY;
ufs_error (sb, "ufs_read_cylinder", "can't read cylinder group block %u", cgno); ufs_error (sb, "ufs_read_cylinder", "can't read cylinder group block %u", cgno);
return false;
} }
/* /*
@ -156,15 +159,14 @@ struct ufs_cg_private_info * ufs_load_cylinder (
UFSD("EXIT (FAILED)\n"); UFSD("EXIT (FAILED)\n");
return NULL; return NULL;
} }
else {
UFSD("EXIT\n");
return sbi->s_ucpi[cgno];
}
} else { } else {
ufs_read_cylinder (sb, cgno, cgno); if (unlikely(!ufs_read_cylinder (sb, cgno, cgno))) {
UFSD("EXIT\n"); UFSD("EXIT (FAILED)\n");
return sbi->s_ucpi[cgno]; return NULL;
}
} }
UFSD("EXIT\n");
return sbi->s_ucpi[cgno];
} }
/* /*
* Cylinder group number cg is in cache but it was not last used, * Cylinder group number cg is in cache but it was not last used,
@ -195,7 +197,10 @@ struct ufs_cg_private_info * ufs_load_cylinder (
sbi->s_ucpi[j] = sbi->s_ucpi[j-1]; sbi->s_ucpi[j] = sbi->s_ucpi[j-1];
} }
sbi->s_ucpi[0] = ucpi; sbi->s_ucpi[0] = ucpi;
ufs_read_cylinder (sb, cgno, 0); if (unlikely(!ufs_read_cylinder (sb, cgno, 0))) {
UFSD("EXIT (FAILED)\n");
return NULL;
}
} }
UFSD("EXIT\n"); UFSD("EXIT\n");
return sbi->s_ucpi[0]; return sbi->s_ucpi[0];

View file

@ -81,10 +81,9 @@ ino_t ufs_inode_by_name(struct inode *dir, const struct qstr *qstr)
} }
/* Releases the page */ int ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de, struct folio *folio, struct inode *inode,
struct folio *folio, struct inode *inode, bool update_times)
bool update_times)
{ {
loff_t pos = folio_pos(folio) + offset_in_folio(folio, de); loff_t pos = folio_pos(folio) + offset_in_folio(folio, de);
unsigned len = fs16_to_cpu(dir->i_sb, de->d_reclen); unsigned len = fs16_to_cpu(dir->i_sb, de->d_reclen);
@ -92,17 +91,19 @@ void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
folio_lock(folio); folio_lock(folio);
err = ufs_prepare_chunk(folio, pos, len); err = ufs_prepare_chunk(folio, pos, len);
BUG_ON(err); if (unlikely(err)) {
folio_unlock(folio);
return err;
}
de->d_ino = cpu_to_fs32(dir->i_sb, inode->i_ino); de->d_ino = cpu_to_fs32(dir->i_sb, inode->i_ino);
ufs_set_de_type(dir->i_sb, de, inode->i_mode); ufs_set_de_type(dir->i_sb, de, inode->i_mode);
ufs_commit_chunk(folio, pos, len); ufs_commit_chunk(folio, pos, len);
folio_release_kmap(folio, de);
if (update_times) if (update_times)
inode_set_mtime_to_ts(dir, inode_set_ctime_current(dir)); inode_set_mtime_to_ts(dir, inode_set_ctime_current(dir));
mark_inode_dirty(dir); mark_inode_dirty(dir);
ufs_handle_dirsync(dir); return ufs_handle_dirsync(dir);
} }
static bool ufs_check_folio(struct folio *folio, char *kaddr) static bool ufs_check_folio(struct folio *folio, char *kaddr)
@ -505,8 +506,7 @@ int ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir,
if (de->d_reclen == 0) { if (de->d_reclen == 0) {
ufs_error(inode->i_sb, __func__, ufs_error(inode->i_sb, __func__,
"zero-length directory entry"); "zero-length directory entry");
err = -EIO; return -EIO;
goto out;
} }
pde = de; pde = de;
de = ufs_next_entry(sb, de); de = ufs_next_entry(sb, de);
@ -516,18 +516,17 @@ int ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir,
pos = folio_pos(folio) + from; pos = folio_pos(folio) + from;
folio_lock(folio); folio_lock(folio);
err = ufs_prepare_chunk(folio, pos, to - from); err = ufs_prepare_chunk(folio, pos, to - from);
BUG_ON(err); if (unlikely(err)) {
folio_unlock(folio);
return err;
}
if (pde) if (pde)
pde->d_reclen = cpu_to_fs16(sb, to - from); pde->d_reclen = cpu_to_fs16(sb, to - from);
dir->d_ino = 0; dir->d_ino = 0;
ufs_commit_chunk(folio, pos, to - from); ufs_commit_chunk(folio, pos, to - from);
inode_set_mtime_to_ts(inode, inode_set_ctime_current(inode)); inode_set_mtime_to_ts(inode, inode_set_ctime_current(inode));
mark_inode_dirty(inode); mark_inode_dirty(inode);
err = ufs_handle_dirsync(inode); return ufs_handle_dirsync(inode);
out:
folio_release_kmap(folio, kaddr);
UFSD("EXIT\n");
return err;
} }
int ufs_make_empty(struct inode * inode, struct inode *dir) int ufs_make_empty(struct inode * inode, struct inode *dir)

View file

@ -42,4 +42,5 @@ const struct file_operations ufs_file_operations = {
.open = generic_file_open, .open = generic_file_open,
.fsync = generic_file_fsync, .fsync = generic_file_fsync,
.splice_read = filemap_splice_read, .splice_read = filemap_splice_read,
.splice_write = iter_file_splice_write,
}; };

View file

@ -220,7 +220,7 @@ static u64 ufs_frag_map(struct inode *inode, unsigned offsets[4], int depth)
*/ */
static bool static bool
ufs_extend_tail(struct inode *inode, u64 writes_to, ufs_extend_tail(struct inode *inode, u64 writes_to,
int *err, struct page *locked_page) int *err, struct folio *locked_folio)
{ {
struct ufs_inode_info *ufsi = UFS_I(inode); struct ufs_inode_info *ufsi = UFS_I(inode);
struct super_block *sb = inode->i_sb; struct super_block *sb = inode->i_sb;
@ -239,7 +239,7 @@ ufs_extend_tail(struct inode *inode, u64 writes_to,
p = ufs_get_direct_data_ptr(uspi, ufsi, block); p = ufs_get_direct_data_ptr(uspi, ufsi, block);
tmp = ufs_new_fragments(inode, p, lastfrag, ufs_data_ptr_to_cpu(sb, p), tmp = ufs_new_fragments(inode, p, lastfrag, ufs_data_ptr_to_cpu(sb, p),
new_size - (lastfrag & uspi->s_fpbmask), err, new_size - (lastfrag & uspi->s_fpbmask), err,
locked_page); locked_folio);
return tmp != 0; return tmp != 0;
} }
@ -250,12 +250,11 @@ ufs_extend_tail(struct inode *inode, u64 writes_to,
* @new_fragment: number of new allocated fragment(s) * @new_fragment: number of new allocated fragment(s)
* @err: we set it if something wrong * @err: we set it if something wrong
* @new: we set it if we allocate new block * @new: we set it if we allocate new block
* @locked_page: for ufs_new_fragments() * @locked_folio: for ufs_new_fragments()
*/ */
static u64 static u64 ufs_inode_getfrag(struct inode *inode, unsigned index,
ufs_inode_getfrag(struct inode *inode, unsigned index,
sector_t new_fragment, int *err, sector_t new_fragment, int *err,
int *new, struct page *locked_page) int *new, struct folio *locked_folio)
{ {
struct ufs_inode_info *ufsi = UFS_I(inode); struct ufs_inode_info *ufsi = UFS_I(inode);
struct super_block *sb = inode->i_sb; struct super_block *sb = inode->i_sb;
@ -264,11 +263,6 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
unsigned nfrags = uspi->s_fpb; unsigned nfrags = uspi->s_fpb;
void *p; void *p;
/* TODO : to be done for write support
if ( (flags & UFS_TYPE_MASK) == UFS_TYPE_UFS2)
goto ufs2;
*/
p = ufs_get_direct_data_ptr(uspi, ufsi, index); p = ufs_get_direct_data_ptr(uspi, ufsi, index);
tmp = ufs_data_ptr_to_cpu(sb, p); tmp = ufs_data_ptr_to_cpu(sb, p);
if (tmp) if (tmp)
@ -288,7 +282,7 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
goal += uspi->s_fpb; goal += uspi->s_fpb;
} }
tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment), tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment),
goal, nfrags, err, locked_page); goal, nfrags, err, locked_folio);
if (!tmp) { if (!tmp) {
*err = -ENOSPC; *err = -ENOSPC;
@ -303,21 +297,6 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
mark_inode_dirty(inode); mark_inode_dirty(inode);
out: out:
return tmp + uspi->s_sbbase; return tmp + uspi->s_sbbase;
/* This part : To be implemented ....
Required only for writing, not required for READ-ONLY.
ufs2:
u2_block = ufs_fragstoblks(fragment);
u2_blockoff = ufs_fragnum(fragment);
p = ufsi->i_u1.u2_i_data + block;
goal = 0;
repeat2:
tmp = fs32_to_cpu(sb, *p);
lastfrag = ufsi->i_lastfrag;
*/
} }
/** /**
@ -329,12 +308,11 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
* (block will hold this fragment and also uspi->s_fpb-1) * (block will hold this fragment and also uspi->s_fpb-1)
* @err: see ufs_inode_getfrag() * @err: see ufs_inode_getfrag()
* @new: see ufs_inode_getfrag() * @new: see ufs_inode_getfrag()
* @locked_page: see ufs_inode_getfrag() * @locked_folio: see ufs_inode_getfrag()
*/ */
static u64 static u64 ufs_inode_getblock(struct inode *inode, u64 ind_block,
ufs_inode_getblock(struct inode *inode, u64 ind_block, unsigned index, sector_t new_fragment, int *err,
unsigned index, sector_t new_fragment, int *err, int *new, struct folio *locked_folio)
int *new, struct page *locked_page)
{ {
struct super_block *sb = inode->i_sb; struct super_block *sb = inode->i_sb;
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi; struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
@ -369,7 +347,7 @@ ufs_inode_getblock(struct inode *inode, u64 ind_block,
else else
goal = bh->b_blocknr + uspi->s_fpb; goal = bh->b_blocknr + uspi->s_fpb;
tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment), goal, tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment), goal,
uspi->s_fpb, err, locked_page); uspi->s_fpb, err, locked_folio);
if (!tmp) if (!tmp)
goto out; goto out;
@ -434,14 +412,14 @@ static int ufs_getfrag_block(struct inode *inode, sector_t fragment, struct buff
unsigned tailfrags = lastfrag & uspi->s_fpbmask; unsigned tailfrags = lastfrag & uspi->s_fpbmask;
if (tailfrags && fragment >= lastfrag) { if (tailfrags && fragment >= lastfrag) {
if (!ufs_extend_tail(inode, fragment, if (!ufs_extend_tail(inode, fragment,
&err, bh_result->b_page)) &err, bh_result->b_folio))
goto out; goto out;
} }
} }
if (depth == 1) { if (depth == 1) {
phys64 = ufs_inode_getfrag(inode, offsets[0], fragment, phys64 = ufs_inode_getfrag(inode, offsets[0], fragment,
&err, &new, bh_result->b_page); &err, &new, bh_result->b_folio);
} else { } else {
int i; int i;
phys64 = ufs_inode_getfrag(inode, offsets[0], fragment, phys64 = ufs_inode_getfrag(inode, offsets[0], fragment,
@ -450,7 +428,7 @@ static int ufs_getfrag_block(struct inode *inode, sector_t fragment, struct buff
phys64 = ufs_inode_getblock(inode, phys64, offsets[i], phys64 = ufs_inode_getblock(inode, phys64, offsets[i],
fragment, &err, NULL, NULL); fragment, &err, NULL, NULL);
phys64 = ufs_inode_getblock(inode, phys64, offsets[depth - 1], phys64 = ufs_inode_getblock(inode, phys64, offsets[depth - 1],
fragment, &err, &new, bh_result->b_page); fragment, &err, &new, bh_result->b_folio);
} }
out: out:
if (phys64) { if (phys64) {
@ -898,91 +876,84 @@ static inline void free_data(struct to_free *ctx, u64 from, unsigned count)
#define DIRECT_FRAGMENT ((inode->i_size + uspi->s_fsize - 1) >> uspi->s_fshift) #define DIRECT_FRAGMENT ((inode->i_size + uspi->s_fsize - 1) >> uspi->s_fshift)
/*
* used only for truncation down to direct blocks.
*/
static void ufs_trunc_direct(struct inode *inode) static void ufs_trunc_direct(struct inode *inode)
{ {
struct ufs_inode_info *ufsi = UFS_I(inode); struct ufs_inode_info *ufsi = UFS_I(inode);
struct super_block * sb; struct super_block *sb = inode->i_sb;
struct ufs_sb_private_info * uspi; struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
void *p; unsigned int new_frags, old_frags;
u64 frag1, frag2, frag3, frag4, block1, block2; unsigned int old_slot, new_slot;
unsigned int old_tail, new_tail;
struct to_free ctx = {.inode = inode}; struct to_free ctx = {.inode = inode};
unsigned i, tmp;
UFSD("ENTER: ino %lu\n", inode->i_ino); UFSD("ENTER: ino %lu\n", inode->i_ino);
sb = inode->i_sb; new_frags = DIRECT_FRAGMENT;
uspi = UFS_SB(sb)->s_uspi; // new_frags = first fragment past the new EOF
old_frags = min_t(u64, UFS_NDIR_FRAGMENT, ufsi->i_lastfrag);
// old_frags = first fragment past the old EOF or covered by indirects
frag1 = DIRECT_FRAGMENT; if (new_frags >= old_frags) // expanding - nothing to free
frag4 = min_t(u64, UFS_NDIR_FRAGMENT, ufsi->i_lastfrag); goto done;
frag2 = ((frag1 & uspi->s_fpbmask) ? ((frag1 | uspi->s_fpbmask) + 1) : frag1);
frag3 = frag4 & ~uspi->s_fpbmask;
block1 = block2 = 0;
if (frag2 > frag3) {
frag2 = frag4;
frag3 = frag4 = 0;
} else if (frag2 < frag3) {
block1 = ufs_fragstoblks (frag2);
block2 = ufs_fragstoblks (frag3);
}
UFSD("ino %lu, frag1 %llu, frag2 %llu, block1 %llu, block2 %llu," old_tail = ufs_fragnum(old_frags);
" frag3 %llu, frag4 %llu\n", inode->i_ino, old_slot = ufs_fragstoblks(old_frags);
(unsigned long long)frag1, (unsigned long long)frag2, new_tail = ufs_fragnum(new_frags);
(unsigned long long)block1, (unsigned long long)block2, new_slot = ufs_fragstoblks(new_frags);
(unsigned long long)frag3, (unsigned long long)frag4);
if (frag1 >= frag2) if (old_slot == new_slot) { // old_tail > 0
goto next1; void *p = ufs_get_direct_data_ptr(uspi, ufsi, old_slot);
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
/*
* Free first free fragments
*/
p = ufs_get_direct_data_ptr(uspi, ufsi, ufs_fragstoblks(frag1));
tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp )
ufs_panic (sb, "ufs_trunc_direct", "internal error");
frag2 -= frag1;
frag1 = ufs_fragnum (frag1);
ufs_free_fragments(inode, tmp + frag1, frag2);
next1:
/*
* Free whole blocks
*/
for (i = block1 ; i < block2; i++) {
p = ufs_get_direct_data_ptr(uspi, ufsi, i);
tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp) if (!tmp)
continue; ufs_panic(sb, __func__, "internal error");
write_seqlock(&ufsi->meta_lock); if (!new_tail) {
ufs_data_ptr_clear(uspi, p); write_seqlock(&ufsi->meta_lock);
write_sequnlock(&ufsi->meta_lock); ufs_data_ptr_clear(uspi, p);
write_sequnlock(&ufsi->meta_lock);
}
ufs_free_fragments(inode, tmp + new_tail, old_tail - new_tail);
} else {
unsigned int slot = new_slot;
free_data(&ctx, tmp, uspi->s_fpb); if (new_tail) {
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot++);
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp)
ufs_panic(sb, __func__, "internal error");
ufs_free_fragments(inode, tmp + new_tail,
uspi->s_fpb - new_tail);
}
while (slot < old_slot) {
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot++);
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp)
continue;
write_seqlock(&ufsi->meta_lock);
ufs_data_ptr_clear(uspi, p);
write_sequnlock(&ufsi->meta_lock);
free_data(&ctx, tmp, uspi->s_fpb);
}
free_data(&ctx, 0, 0);
if (old_tail) {
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot);
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp)
ufs_panic(sb, __func__, "internal error");
write_seqlock(&ufsi->meta_lock);
ufs_data_ptr_clear(uspi, p);
write_sequnlock(&ufsi->meta_lock);
ufs_free_fragments(inode, tmp, old_tail);
}
} }
done:
free_data(&ctx, 0, 0);
if (frag3 >= frag4)
goto next3;
/*
* Free last free fragments
*/
p = ufs_get_direct_data_ptr(uspi, ufsi, ufs_fragstoblks(frag3));
tmp = ufs_data_ptr_to_cpu(sb, p);
if (!tmp )
ufs_panic(sb, "ufs_truncate_direct", "internal error");
frag4 = ufs_fragnum (frag4);
write_seqlock(&ufsi->meta_lock);
ufs_data_ptr_clear(uspi, p);
write_sequnlock(&ufsi->meta_lock);
ufs_free_fragments (inode, tmp, frag4);
next3:
UFSD("EXIT: ino %lu\n", inode->i_ino); UFSD("EXIT: ino %lu\n", inode->i_ino);
} }

View file

@ -210,20 +210,18 @@ static int ufs_unlink(struct inode *dir, struct dentry *dentry)
struct inode * inode = d_inode(dentry); struct inode * inode = d_inode(dentry);
struct ufs_dir_entry *de; struct ufs_dir_entry *de;
struct folio *folio; struct folio *folio;
int err = -ENOENT; int err;
de = ufs_find_entry(dir, &dentry->d_name, &folio); de = ufs_find_entry(dir, &dentry->d_name, &folio);
if (!de) if (!de)
goto out; return -ENOENT;
err = ufs_delete_entry(dir, de, folio); err = ufs_delete_entry(dir, de, folio);
if (err) if (!err) {
goto out; inode_set_ctime_to_ts(inode, inode_get_ctime(dir));
inode_dec_link_count(inode);
inode_set_ctime_to_ts(inode, inode_get_ctime(dir)); }
inode_dec_link_count(inode); folio_release_kmap(folio, de);
err = 0;
out:
return err; return err;
} }
@ -253,14 +251,14 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
struct ufs_dir_entry * dir_de = NULL; struct ufs_dir_entry * dir_de = NULL;
struct folio *old_folio; struct folio *old_folio;
struct ufs_dir_entry *old_de; struct ufs_dir_entry *old_de;
int err = -ENOENT; int err;
if (flags & ~RENAME_NOREPLACE) if (flags & ~RENAME_NOREPLACE)
return -EINVAL; return -EINVAL;
old_de = ufs_find_entry(old_dir, &old_dentry->d_name, &old_folio); old_de = ufs_find_entry(old_dir, &old_dentry->d_name, &old_folio);
if (!old_de) if (!old_de)
goto out; return -ENOENT;
if (S_ISDIR(old_inode->i_mode)) { if (S_ISDIR(old_inode->i_mode)) {
err = -EIO; err = -EIO;
@ -281,7 +279,10 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_folio); new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_folio);
if (!new_de) if (!new_de)
goto out_dir; goto out_dir;
ufs_set_link(new_dir, new_de, new_folio, old_inode, 1); err = ufs_set_link(new_dir, new_de, new_folio, old_inode, 1);
folio_release_kmap(new_folio, new_de);
if (err)
goto out_dir;
inode_set_ctime_current(new_inode); inode_set_ctime_current(new_inode);
if (dir_de) if (dir_de)
drop_nlink(new_inode); drop_nlink(new_inode);
@ -299,26 +300,20 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
* rename. * rename.
*/ */
inode_set_ctime_current(old_inode); inode_set_ctime_current(old_inode);
ufs_delete_entry(old_dir, old_de, old_folio);
mark_inode_dirty(old_inode); mark_inode_dirty(old_inode);
if (dir_de) { err = ufs_delete_entry(old_dir, old_de, old_folio);
if (!err && dir_de) {
if (old_dir != new_dir) if (old_dir != new_dir)
ufs_set_link(old_inode, dir_de, dir_folio, new_dir, 0); err = ufs_set_link(old_inode, dir_de, dir_folio,
else new_dir, 0);
folio_release_kmap(dir_folio, dir_de);
inode_dec_link_count(old_dir); inode_dec_link_count(old_dir);
} }
return 0;
out_dir: out_dir:
if (dir_de) if (dir_de)
folio_release_kmap(dir_folio, dir_de); folio_release_kmap(dir_folio, dir_de);
out_old: out_old:
folio_release_kmap(old_folio, old_de); folio_release_kmap(old_folio, old_de);
out:
return err; return err;
} }

View file

@ -505,7 +505,6 @@ static int ufs_read_cylinder_structures(struct super_block *sb)
{ {
struct ufs_sb_info *sbi = UFS_SB(sb); struct ufs_sb_info *sbi = UFS_SB(sb);
struct ufs_sb_private_info *uspi = sbi->s_uspi; struct ufs_sb_private_info *uspi = sbi->s_uspi;
struct ufs_buffer_head * ubh;
unsigned char * base, * space; unsigned char * base, * space;
unsigned size, blks, i; unsigned size, blks, i;
@ -521,21 +520,13 @@ static int ufs_read_cylinder_structures(struct super_block *sb)
if (!base) if (!base)
goto failed; goto failed;
sbi->s_csp = (struct ufs_csum *)space; sbi->s_csp = (struct ufs_csum *)space;
for (i = 0; i < blks; i += uspi->s_fpb) { for (i = 0; i < blks; i++) {
size = uspi->s_bsize; struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
if (i + uspi->s_fpb > blks) if (!bh)
size = (blks - i) * uspi->s_fsize;
ubh = ubh_bread(sb, uspi->s_csaddr + i, size);
if (!ubh)
goto failed; goto failed;
memcpy(space, bh->b_data, uspi->s_fsize);
ubh_ubhcpymem (space, ubh, size); space += uspi->s_fsize;
brelse (bh);
space += size;
ubh_brelse (ubh);
ubh = NULL;
} }
/* /*
@ -645,7 +636,6 @@ static void ufs_put_super_internal(struct super_block *sb)
{ {
struct ufs_sb_info *sbi = UFS_SB(sb); struct ufs_sb_info *sbi = UFS_SB(sb);
struct ufs_sb_private_info *uspi = sbi->s_uspi; struct ufs_sb_private_info *uspi = sbi->s_uspi;
struct ufs_buffer_head * ubh;
unsigned char * base, * space; unsigned char * base, * space;
unsigned blks, size, i; unsigned blks, size, i;
@ -656,18 +646,17 @@ static void ufs_put_super_internal(struct super_block *sb)
size = uspi->s_cssize; size = uspi->s_cssize;
blks = (size + uspi->s_fsize - 1) >> uspi->s_fshift; blks = (size + uspi->s_fsize - 1) >> uspi->s_fshift;
base = space = (char*) sbi->s_csp; base = space = (char*) sbi->s_csp;
for (i = 0; i < blks; i += uspi->s_fpb) { for (i = 0; i < blks; i++, space += uspi->s_fsize) {
size = uspi->s_bsize; struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
if (i + uspi->s_fpb > blks)
size = (blks - i) * uspi->s_fsize;
ubh = ubh_bread(sb, uspi->s_csaddr + i, size); if (unlikely(!bh)) { // better than an oops...
ufs_panic(sb, __func__,
ubh_memcpyubh (ubh, space, size); "can't write part of cylinder group summary");
space += size; continue;
ubh_mark_buffer_uptodate (ubh, 1); }
ubh_mark_buffer_dirty (ubh); memcpy(bh->b_data, space, uspi->s_fsize);
ubh_brelse (ubh); mark_buffer_dirty(bh);
brelse(bh);
} }
for (i = 0; i < sbi->s_cg_loaded; i++) { for (i = 0; i < sbi->s_cg_loaded; i++) {
ufs_put_cylinder (sb, i); ufs_put_cylinder (sb, i);
@ -1240,11 +1229,7 @@ static int ufs_fill_super(struct super_block *sb, void *data, int silent)
else else
uspi->s_apbshift = uspi->s_bshift - 2; uspi->s_apbshift = uspi->s_bshift - 2;
uspi->s_2apbshift = uspi->s_apbshift * 2;
uspi->s_3apbshift = uspi->s_apbshift * 3;
uspi->s_apb = 1 << uspi->s_apbshift; uspi->s_apb = 1 << uspi->s_apbshift;
uspi->s_2apb = 1 << uspi->s_2apbshift;
uspi->s_3apb = 1 << uspi->s_3apbshift;
uspi->s_apbmask = uspi->s_apb - 1; uspi->s_apbmask = uspi->s_apb - 1;
uspi->s_nspfshift = uspi->s_fshift - UFS_SECTOR_BITS; uspi->s_nspfshift = uspi->s_fshift - UFS_SECTOR_BITS;
uspi->s_nspb = uspi->s_nspf << uspi->s_fpbshift; uspi->s_nspb = uspi->s_nspf << uspi->s_fpbshift;

View file

@ -88,10 +88,10 @@ struct ufs_inode_info {
#endif #endif
/* balloc.c */ /* balloc.c */
extern void ufs_free_fragments (struct inode *, u64, unsigned); void ufs_free_fragments (struct inode *, u64 fragment, unsigned count);
extern void ufs_free_blocks (struct inode *, u64, unsigned); void ufs_free_blocks (struct inode *, u64 fragment, unsigned count);
extern u64 ufs_new_fragments(struct inode *, void *, u64, u64, u64 ufs_new_fragments(struct inode *, void *, u64 fragment, u64 goal,
unsigned, int *, struct page *); unsigned count, int *err, struct folio *);
/* cylinder.c */ /* cylinder.c */
extern struct ufs_cg_private_info * ufs_load_cylinder (struct super_block *, unsigned); extern struct ufs_cg_private_info * ufs_load_cylinder (struct super_block *, unsigned);
@ -108,8 +108,8 @@ struct ufs_dir_entry *ufs_find_entry(struct inode *, const struct qstr *,
int ufs_delete_entry(struct inode *, struct ufs_dir_entry *, struct folio *); int ufs_delete_entry(struct inode *, struct ufs_dir_entry *, struct folio *);
int ufs_empty_dir(struct inode *); int ufs_empty_dir(struct inode *);
struct ufs_dir_entry *ufs_dotdot(struct inode *, struct folio **); struct ufs_dir_entry *ufs_dotdot(struct inode *, struct folio **);
void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de, int ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
struct folio *folio, struct inode *inode, bool update_times); struct folio *folio, struct inode *inode, bool update_times);
/* file.c */ /* file.c */
extern const struct inode_operations ufs_file_inode_operations; extern const struct inode_operations ufs_file_inode_operations;

View file

@ -775,12 +775,8 @@ struct ufs_sb_private_info {
__u32 s_fpbmask; /* fragments per block mask */ __u32 s_fpbmask; /* fragments per block mask */
__u32 s_apb; /* address per block */ __u32 s_apb; /* address per block */
__u32 s_2apb; /* address per block^2 */
__u32 s_3apb; /* address per block^3 */
__u32 s_apbmask; /* address per block mask */ __u32 s_apbmask; /* address per block mask */
__u32 s_apbshift; /* address per block shift */ __u32 s_apbshift; /* address per block shift */
__u32 s_2apbshift; /* address per block shift * 2 */
__u32 s_3apbshift; /* address per block shift * 3 */
__u32 s_nspfshift; /* number of sector per fragment shift */ __u32 s_nspfshift; /* number of sector per fragment shift */
__u32 s_nspb; /* number of sector per block */ __u32 s_nspb; /* number of sector per block */
__u32 s_inopf; /* inodes per fragment */ __u32 s_inopf; /* inodes per fragment */

View file

@ -99,20 +99,6 @@ void ubh_mark_buffer_dirty (struct ufs_buffer_head * ubh)
mark_buffer_dirty (ubh->bh[i]); mark_buffer_dirty (ubh->bh[i]);
} }
void ubh_mark_buffer_uptodate (struct ufs_buffer_head * ubh, int flag)
{
unsigned i;
if (!ubh)
return;
if (flag) {
for ( i = 0; i < ubh->count; i++ )
set_buffer_uptodate (ubh->bh[i]);
} else {
for ( i = 0; i < ubh->count; i++ )
clear_buffer_uptodate (ubh->bh[i]);
}
}
void ubh_sync_block(struct ufs_buffer_head *ubh) void ubh_sync_block(struct ufs_buffer_head *ubh)
{ {
if (ubh) { if (ubh) {
@ -146,38 +132,6 @@ int ubh_buffer_dirty (struct ufs_buffer_head * ubh)
return result; return result;
} }
void _ubh_ubhcpymem_(struct ufs_sb_private_info * uspi,
unsigned char * mem, struct ufs_buffer_head * ubh, unsigned size)
{
unsigned len, bhno;
if (size > (ubh->count << uspi->s_fshift))
size = ubh->count << uspi->s_fshift;
bhno = 0;
while (size) {
len = min_t(unsigned int, size, uspi->s_fsize);
memcpy (mem, ubh->bh[bhno]->b_data, len);
mem += uspi->s_fsize;
size -= len;
bhno++;
}
}
void _ubh_memcpyubh_(struct ufs_sb_private_info * uspi,
struct ufs_buffer_head * ubh, unsigned char * mem, unsigned size)
{
unsigned len, bhno;
if (size > (ubh->count << uspi->s_fshift))
size = ubh->count << uspi->s_fshift;
bhno = 0;
while (size) {
len = min_t(unsigned int, size, uspi->s_fsize);
memcpy (ubh->bh[bhno]->b_data, mem, len);
mem += uspi->s_fsize;
size -= len;
bhno++;
}
}
dev_t dev_t
ufs_get_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi) ufs_get_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi)
{ {

View file

@ -263,14 +263,9 @@ extern struct ufs_buffer_head * ubh_bread_uspi(struct ufs_sb_private_info *, str
extern void ubh_brelse (struct ufs_buffer_head *); extern void ubh_brelse (struct ufs_buffer_head *);
extern void ubh_brelse_uspi (struct ufs_sb_private_info *); extern void ubh_brelse_uspi (struct ufs_sb_private_info *);
extern void ubh_mark_buffer_dirty (struct ufs_buffer_head *); extern void ubh_mark_buffer_dirty (struct ufs_buffer_head *);
extern void ubh_mark_buffer_uptodate (struct ufs_buffer_head *, int);
extern void ubh_sync_block(struct ufs_buffer_head *); extern void ubh_sync_block(struct ufs_buffer_head *);
extern void ubh_bforget (struct ufs_buffer_head *); extern void ubh_bforget (struct ufs_buffer_head *);
extern int ubh_buffer_dirty (struct ufs_buffer_head *); extern int ubh_buffer_dirty (struct ufs_buffer_head *);
#define ubh_ubhcpymem(mem,ubh,size) _ubh_ubhcpymem_(uspi,mem,ubh,size)
extern void _ubh_ubhcpymem_(struct ufs_sb_private_info *, unsigned char *, struct ufs_buffer_head *, unsigned);
#define ubh_memcpyubh(ubh,mem,size) _ubh_memcpyubh_(uspi,ubh,mem,size)
extern void _ubh_memcpyubh_(struct ufs_sb_private_info *, struct ufs_buffer_head *, unsigned char *, unsigned);
/* This functions works with cache pages*/ /* This functions works with cache pages*/
struct folio *ufs_get_locked_folio(struct address_space *mapping, pgoff_t index); struct folio *ufs_get_locked_folio(struct address_space *mapping, pgoff_t index);
@ -455,65 +450,69 @@ static inline unsigned _ubh_find_last_zero_bit_(
return (base << uspi->s_bpfshift) + pos - begin; return (base << uspi->s_bpfshift) + pos - begin;
} }
#define ubh_isblockclear(ubh,begin,block) (!_ubh_isblockset_(uspi,ubh,begin,block)) static inline int ubh_isblockset(struct ufs_sb_private_info *uspi,
struct ufs_cg_private_info *ucpi, unsigned int frag)
#define ubh_isblockset(ubh,begin,block) _ubh_isblockset_(uspi,ubh,begin,block)
static inline int _ubh_isblockset_(struct ufs_sb_private_info * uspi,
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
{ {
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
u8 mask; u8 mask;
switch (uspi->s_fpb) { switch (uspi->s_fpb) {
case 8: case 8:
return (*ubh_get_addr (ubh, begin + block) == 0xff); return *p == 0xff;
case 4: case 4:
mask = 0x0f << ((block & 0x01) << 2); mask = 0x0f << (frag & 4);
return (*ubh_get_addr (ubh, begin + (block >> 1)) & mask) == mask; return (*p & mask) == mask;
case 2: case 2:
mask = 0x03 << ((block & 0x03) << 1); mask = 0x03 << (frag & 6);
return (*ubh_get_addr (ubh, begin + (block >> 2)) & mask) == mask; return (*p & mask) == mask;
case 1: case 1:
mask = 0x01 << (block & 0x07); mask = 0x01 << (frag & 7);
return (*ubh_get_addr (ubh, begin + (block >> 3)) & mask) == mask; return (*p & mask) == mask;
} }
return 0; return 0;
} }
#define ubh_clrblock(ubh,begin,block) _ubh_clrblock_(uspi,ubh,begin,block) static inline void ubh_clrblock(struct ufs_sb_private_info *uspi,
static inline void _ubh_clrblock_(struct ufs_sb_private_info * uspi, struct ufs_cg_private_info *ucpi, unsigned int frag)
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
{ {
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
switch (uspi->s_fpb) { switch (uspi->s_fpb) {
case 8: case 8:
*ubh_get_addr (ubh, begin + block) = 0x00; *p = 0x00;
return; return;
case 4: case 4:
*ubh_get_addr (ubh, begin + (block >> 1)) &= ~(0x0f << ((block & 0x01) << 2)); *p &= ~(0x0f << (frag & 4));
return; return;
case 2: case 2:
*ubh_get_addr (ubh, begin + (block >> 2)) &= ~(0x03 << ((block & 0x03) << 1)); *p &= ~(0x03 << (frag & 6));
return; return;
case 1: case 1:
*ubh_get_addr (ubh, begin + (block >> 3)) &= ~(0x01 << ((block & 0x07))); *p &= ~(0x01 << (frag & 7));
return; return;
} }
} }
#define ubh_setblock(ubh,begin,block) _ubh_setblock_(uspi,ubh,begin,block) static inline void ubh_setblock(struct ufs_sb_private_info * uspi,
static inline void _ubh_setblock_(struct ufs_sb_private_info * uspi, struct ufs_cg_private_info *ucpi, unsigned int frag)
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
{ {
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
switch (uspi->s_fpb) { switch (uspi->s_fpb) {
case 8: case 8:
*ubh_get_addr(ubh, begin + block) = 0xff; *p = 0xff;
return; return;
case 4: case 4:
*ubh_get_addr(ubh, begin + (block >> 1)) |= (0x0f << ((block & 0x01) << 2)); *p |= 0x0f << (frag & 4);
return; return;
case 2: case 2:
*ubh_get_addr(ubh, begin + (block >> 2)) |= (0x03 << ((block & 0x03) << 1)); *p |= 0x03 << (frag & 6);
return; return;
case 1: case 1:
*ubh_get_addr(ubh, begin + (block >> 3)) |= (0x01 << ((block & 0x07))); *p |= 0x01 << (frag & 7);
return; return;
} }
} }