if (adapter->fw_done_rc) {
                dev_err(dev, "Couldn't map LTB, rc = %d\n",
                        adapter->fw_done_rc);
-               rc = -1;
+               rc = -EIO;
                goto out;
        }
        rc = 0;
 {
        struct device *dev = &adapter->vdev->dev;
        dma_addr_t stok;
+       int rc;
 
        stok = dma_map_single(dev, &adapter->stats,
                              sizeof(struct ibmvnic_statistics),
                              DMA_FROM_DEVICE);
-       if (dma_mapping_error(dev, stok)) {
-               dev_err(dev, "Couldn't map stats buffer\n");
-               return -1;
+       rc = dma_mapping_error(dev, stok);
+       if (rc) {
+               dev_err(dev, "Couldn't map stats buffer, rc = %d\n", rc);
+               return rc;
        }
 
        adapter->stats_token = stok;
        u64 num_pools;
        u64 pool_size;          /* # of buffers in one pool */
        u64 buff_size;
-       int i, j;
+       int i, j, rc;
 
        pool_size = adapter->req_rx_add_entries_per_subcrq;
        num_pools = adapter->req_rx_queues;
                                   GFP_KERNEL);
        if (!adapter->rx_pool) {
                dev_err(dev, "Failed to allocate rx pools\n");
-               return -1;
+               return -ENOMEM;
        }
 
        /* Set num_active_rx_pools early. If we fail below after partial
                                            GFP_KERNEL);
                if (!rx_pool->free_map) {
                        dev_err(dev, "Couldn't alloc free_map %d\n", i);
+                       rc = -ENOMEM;
                        goto out_release;
                }
 
                                           GFP_KERNEL);
                if (!rx_pool->rx_buff) {
                        dev_err(dev, "Couldn't alloc rx buffers\n");
+                       rc = -ENOMEM;
                        goto out_release;
                }
        }
                dev_dbg(dev, "Updating LTB for rx pool %d [%d, %d]\n",
                        i, rx_pool->size, rx_pool->buff_size);
 
-               if (alloc_long_term_buff(adapter, &rx_pool->long_term_buff,
-                                        rx_pool->size * rx_pool->buff_size))
+               rc = alloc_long_term_buff(adapter, &rx_pool->long_term_buff,
+                                         rx_pool->size * rx_pool->buff_size);
+               if (rc)
                        goto out;
 
                for (j = 0; j < rx_pool->size; ++j) {
        /* We failed to allocate one or more LTBs or map them on the VIOS.
         * Hold onto the pools and any LTBs that we did allocate/map.
         */
-       return -1;
+       return rc;
 }
 
 static void release_vpd_data(struct ibmvnic_adapter *adapter)
                                   sizeof(struct ibmvnic_tx_buff),
                                   GFP_KERNEL);
        if (!tx_pool->tx_buff)
-               return -1;
+               return -ENOMEM;
 
        tx_pool->free_map = kcalloc(pool_size, sizeof(int), GFP_KERNEL);
        if (!tx_pool->free_map) {
                kfree(tx_pool->tx_buff);
                tx_pool->tx_buff = NULL;
-               return -1;
+               return -ENOMEM;
        }
 
        for (i = 0; i < pool_size; i++)
        adapter->tx_pool = kcalloc(num_pools,
                                   sizeof(struct ibmvnic_tx_pool), GFP_KERNEL);
        if (!adapter->tx_pool)
-               return -1;
+               return -ENOMEM;
 
        adapter->tso_pool = kcalloc(num_pools,
                                    sizeof(struct ibmvnic_tx_pool), GFP_KERNEL);
        if (!adapter->tso_pool) {
                kfree(adapter->tx_pool);
                adapter->tx_pool = NULL;
-               return -1;
+               return -ENOMEM;
        }
 
        /* Set num_active_tx_pools early. If we fail below after partial
                retry = false;
                if (retry_count > retries) {
                        netdev_warn(netdev, "Login attempts exceeded\n");
-                       return -1;
+                       return -EACCES;
                }
 
                adapter->init_done_rc = 0;
                                                         timeout)) {
                                netdev_warn(netdev,
                                            "Capabilities query timed out\n");
-                               return -1;
+                               return -ETIMEDOUT;
                        }
 
                        rc = init_sub_crqs(adapter);
                        if (rc) {
                                netdev_warn(netdev,
                                            "SCRQ initialization failed\n");
-                               return -1;
+                               return rc;
                        }
 
                        rc = init_sub_crq_irqs(adapter);
                        if (rc) {
                                netdev_warn(netdev,
                                            "SCRQ irq initialization failed\n");
-                               return -1;
+                               return rc;
                        }
                } else if (adapter->init_done_rc) {
-                       netdev_warn(netdev, "Adapter login failed\n");
-                       return -1;
+                       netdev_warn(netdev, "Adapter login failed, init_done_rc = %d\n",
+                                   adapter->init_done_rc);
+                       return -EIO;
                }
        } while (retry);
 
                if (!wait_for_completion_timeout(&adapter->init_done,
                                                 timeout)) {
                        netdev_err(netdev, "timeout setting link state\n");
-                       return -1;
+                       return -ETIMEDOUT;
                }
 
                if (adapter->init_done_rc == PARTIALSUCCESS) {
                                /* If someone else changed the adapter state
                                 * when we dropped the rtnl, fail the reset
                                 */
-                               rc = -1;
+                               rc = -EAGAIN;
                                goto out;
                        }
                        adapter->state = VNIC_CLOSED;
                }
 
                rc = ibmvnic_reset_init(adapter, true);
-               if (rc) {
-                       rc = IBMVNIC_INIT_FAILED;
+               if (rc)
                        goto out;
-               }
 
                /* If the adapter was in PROBE or DOWN state prior to the reset,
                 * exit here.
 
        allqueues = kcalloc(total_queues, sizeof(*allqueues), GFP_KERNEL);
        if (!allqueues)
-               return -1;
+               return -ENOMEM;
 
        for (i = 0; i < total_queues; i++) {
                allqueues[i] = init_sub_crq_queue(adapter);
        for (i = 0; i < registered_queues; i++)
                release_sub_crq_queue(adapter, allqueues[i], 1);
        kfree(allqueues);
-       return -1;
+       return -ENOMEM;
 }
 
 static void send_request_cap(struct ibmvnic_adapter *adapter, int retry)
        if (!adapter->tx_scrq || !adapter->rx_scrq) {
                netdev_err(adapter->netdev,
                           "RX or TX queues are not allocated, device login failed\n");
-               return -1;
+               return -ENOMEM;
        }
 
        release_login_buffer(adapter);
        kfree(login_buffer);
        adapter->login_buf = NULL;
 buf_alloc_failed:
-       return -1;
+       return -ENOMEM;
 }
 
 static int send_request_map(struct ibmvnic_adapter *adapter, dma_addr_t addr,
 
        if (!wait_for_completion_timeout(&adapter->init_done, timeout)) {
                dev_err(dev, "Initialization sequence timed out\n");
-               return -1;
+               return -ETIMEDOUT;
        }
 
        if (adapter->init_done_rc) {
        if (adapter->from_passive_init) {
                adapter->state = VNIC_OPEN;
                adapter->from_passive_init = false;
-               return -1;
+               return -EINVAL;
        }
 
        if (reset &&