return done;
 }
 
-static const int frag_sizes[] = {
-       FRAG_SZ0,
-       FRAG_SZ1,
-       FRAG_SZ2,
-       FRAG_SZ3
-};
-
 void mlx4_en_calc_rx_buf(struct net_device *dev)
 {
        struct mlx4_en_priv *priv = netdev_priv(dev);
                int buf_size = 0;
 
                while (buf_size < eff_mtu) {
-                       priv->frag_info[i].frag_size =
-                               (eff_mtu > buf_size + frag_sizes[i]) ?
-                                       frag_sizes[i] : eff_mtu - buf_size;
-                       priv->frag_info[i].frag_stride =
-                               ALIGN(priv->frag_info[i].frag_size,
-                                     SMP_CACHE_BYTES);
-                       buf_size += priv->frag_info[i].frag_size;
+                       int frag_size = eff_mtu - buf_size;
+
+                       if (i < MLX4_EN_MAX_RX_FRAGS - 1)
+                               frag_size = min(frag_size, 2048);
+
+                       priv->frag_info[i].frag_size = frag_size;
+
+                       priv->frag_info[i].frag_stride = ALIGN(frag_size,
+                                                              SMP_CACHE_BYTES);
+                       buf_size += frag_size;
                        i++;
                }
                priv->rx_page_order = MLX4_EN_ALLOC_PREFER_ORDER;
 
 #define MLX4_EN_ALLOC_PREFER_ORDER min_t(int, get_order(32768),                \
                                         PAGE_ALLOC_COSTLY_ORDER)
 
-/* Receive fragment sizes; we use at most 3 fragments (for 9600 byte MTU
- * and 4K allocations) */
-enum {
-       FRAG_SZ0 = 1536 - NET_IP_ALIGN,
-       FRAG_SZ1 = 4096,
-       FRAG_SZ2 = 4096,
-       FRAG_SZ3 = MLX4_EN_ALLOC_SIZE
-};
 #define MLX4_EN_MAX_RX_FRAGS   4
 
 /* Maximum ring sizes */