@@ -56,7 +56,7 @@ static void
5656create_mkey_callback (int status , struct mlx5_async_work * context );
5757static struct mlx5_ib_mr * reg_create (struct ib_pd * pd , struct ib_umem * umem ,
5858 u64 iova , int access_flags ,
59- unsigned int page_size , bool populate ,
59+ unsigned long page_size , bool populate ,
6060 int access_mode );
6161static int __mlx5_ib_dereg_mr (struct ib_mr * ibmr );
6262
@@ -1115,7 +1115,7 @@ static struct mlx5_ib_mr *alloc_cacheable_mr(struct ib_pd *pd,
11151115 struct mlx5r_cache_rb_key rb_key = {};
11161116 struct mlx5_cache_ent * ent ;
11171117 struct mlx5_ib_mr * mr ;
1118- unsigned int page_size ;
1118+ unsigned long page_size ;
11191119
11201120 if (umem -> is_dmabuf )
11211121 page_size = mlx5_umem_dmabuf_default_pgsz (umem , iova );
@@ -1219,7 +1219,7 @@ reg_create_crossing_vhca_mr(struct ib_pd *pd, u64 iova, u64 length, int access_f
12191219 */
12201220static struct mlx5_ib_mr * reg_create (struct ib_pd * pd , struct ib_umem * umem ,
12211221 u64 iova , int access_flags ,
1222- unsigned int page_size , bool populate ,
1222+ unsigned long page_size , bool populate ,
12231223 int access_mode )
12241224{
12251225 struct mlx5_ib_dev * dev = to_mdev (pd -> device );
@@ -1425,7 +1425,7 @@ static struct ib_mr *create_real_mr(struct ib_pd *pd, struct ib_umem *umem,
14251425 mr = alloc_cacheable_mr (pd , umem , iova , access_flags ,
14261426 MLX5_MKC_ACCESS_MODE_MTT );
14271427 } else {
1428- unsigned int page_size =
1428+ unsigned long page_size =
14291429 mlx5_umem_mkc_find_best_pgsz (dev , umem , iova );
14301430
14311431 mutex_lock (& dev -> slow_path_mutex );
0 commit comments