ino_tree_node_t *irec;
ino_tree_node_t *cur_irec;
int blks_per_cluster;
- int inos_per_cluster;
xfs_agblock_t bno;
int i;
int err;
blks_per_cluster = XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_blocklog;
if (blks_per_cluster == 0)
blks_per_cluster = 1;
- inos_per_cluster = blks_per_cluster * mp->m_sb.sb_inopblock;
for (i = 0; i < PF_THREAD_COUNT; i++) {
err = pthread_create(&args->io_threads[i], NULL,
(cur_irec->ino_isa_dir != 0) ?
B_DIR_INODE : B_INODE);
bno += blks_per_cluster;
- num_inos += inos_per_cluster;
+ num_inos += inodes_per_cluster;
} while (num_inos < XFS_IALLOC_INODES(mp));
}
prefetch_args_t *prev_args)
{
prefetch_args_t *args;
+ long max_queue;
if (!do_prefetch || agno >= mp->m_sb.sb_agcount)
return NULL;
* and not any other associated metadata like directories
*/
- sem_init(&args->ra_count, 0, libxfs_bcache->c_maxcount / thread_count /
- (XFS_IALLOC_BLOCKS(mp) / (XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_blocklog)) / 8);
+ max_queue = libxfs_bcache->c_maxcount / thread_count / 8;
+ if (XFS_INODE_CLUSTER_SIZE(mp) > mp->m_sb.sb_blocksize)
+ max_queue = max_queue * (XFS_INODE_CLUSTER_SIZE(mp) >>
+ mp->m_sb.sb_blocklog) / XFS_IALLOC_BLOCKS(mp);
+
+ sem_init(&args->ra_count, 0, max_queue);
if (!prev_args) {
if (!pf_create_prefetch_thread(args))
chunks_pblock = mp->m_sb.sb_inopblock / XFS_INODES_PER_CHUNK;
max_symlink_blocks = howmany(MAXPATHLEN - 1, mp->m_sb.sb_blocksize);
- inodes_per_cluster = XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog;
+ inodes_per_cluster = MAX(mp->m_sb.sb_inopblock,
+ XFS_INODE_CLUSTER_SIZE(mp) >> mp->m_sb.sb_inodelog);
if (ag_stride) {
thread_count = (glob_agcount + ag_stride - 1) / ag_stride;