Signed-off-by: Bradley Grove <bgr...@attotech.com>
---
 drivers/scsi/esas2r/esas2r_disc.c | 1192 +++++++++++++++++++++++++++++++++++++
 drivers/scsi/esas2r/esas2r_log.c  |  255 ++++++++
 drivers/scsi/esas2r/esas2r_log.h  |  118 ++++
 3 files changed, 1565 insertions(+)
 create mode 100644 drivers/scsi/esas2r/esas2r_disc.c
 create mode 100644 drivers/scsi/esas2r/esas2r_log.c
 create mode 100644 drivers/scsi/esas2r/esas2r_log.h

diff --git a/drivers/scsi/esas2r/esas2r_disc.c 
b/drivers/scsi/esas2r/esas2r_disc.c
new file mode 100644
index 0000000..6eae3c1
--- /dev/null
+++ b/drivers/scsi/esas2r/esas2r_disc.c
@@ -0,0 +1,1191 @@
+/*
+ *  linux/drivers/scsi/esas2r/esas2r_disc.c
+ *      esas2r device discovery routines
+ *
+ *  Copyright (c) 2001-2013 ATTO Technology, Inc.
+ *  (mailto:linuxdriv...@attotech.com)
+ */
+/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
+/*
+ *  This program is free software; you can redistribute it and/or modify
+ *  it under the terms of the GNU General Public License as published by
+ *  the Free Software Foundation; version 2 of the License.
+ *
+ *  This program is distributed in the hope that it will be useful,
+ *  but WITHOUT ANY WARRANTY; without even the implied warranty of
+ *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ *  GNU General Public License for more details.
+ *
+ *  NO WARRANTY
+ *  THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
+ *  CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
+ *  LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
+ *  MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
+ *  solely responsible for determining the appropriateness of using and
+ *  distributing the Program and assumes all risks associated with its
+ *  exercise of rights under this Agreement, including but not limited to
+ *  the risks and costs of program errors, damage to or loss of data,
+ *  programs or equipment, and unavailability or interruption of operations.
+ *
+ *  DISCLAIMER OF LIABILITY
+ *  NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
+ *  DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ *  DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
+ *  ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
+ *  TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
+ *  USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
+ *  HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
+ *
+ *  You should have received a copy of the GNU General Public License
+ *  along with this program; if not, write to the Free Software
+ *  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
+ */
+/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
+
+#include "esas2r.h"
+
+/* Miscellaneous internal discovery routines */
+static void esas2r_disc_abort(struct esas2r_adapter *a,
+                             struct esas2r_request *rq);
+static bool esas2r_disc_continue(struct esas2r_adapter *a,
+                                struct esas2r_request *rq);
+static void esas2r_disc_fix_curr_requests(struct esas2r_adapter *a);
+static u32 esas2r_disc_get_phys_addr(struct esas2r_sg_context *sgc, u64 *addr);
+static bool esas2r_disc_start_request(struct esas2r_adapter *a,
+                                     struct esas2r_request *rq);
+
+/* Internal discovery routines that process the states */
+static bool esas2r_disc_block_dev_scan(struct esas2r_adapter *a,
+                                      struct esas2r_request *rq);
+static void esas2r_disc_block_dev_scan_cb(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq);
+static bool esas2r_disc_dev_add(struct esas2r_adapter *a,
+                               struct esas2r_request *rq);
+static bool esas2r_disc_dev_remove(struct esas2r_adapter *a,
+                                  struct esas2r_request *rq);
+static bool esas2r_disc_part_info(struct esas2r_adapter *a,
+                                 struct esas2r_request *rq);
+static void esas2r_disc_part_info_cb(struct esas2r_adapter *a,
+                                    struct esas2r_request *rq);
+static bool esas2r_disc_passthru_dev_info(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq);
+static void esas2r_disc_passthru_dev_info_cb(struct esas2r_adapter *a,
+                                            struct esas2r_request *rq);
+static bool esas2r_disc_passthru_dev_addr(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq);
+static void esas2r_disc_passthru_dev_addr_cb(struct esas2r_adapter *a,
+                                            struct esas2r_request *rq);
+static bool esas2r_disc_raid_grp_info(struct esas2r_adapter *a,
+                                     struct esas2r_request *rq);
+static void esas2r_disc_raid_grp_info_cb(struct esas2r_adapter *a,
+                                        struct esas2r_request *rq);
+
+void esas2r_disc_initialize(struct esas2r_adapter *a)
+{
+       struct esas2r_sas_nvram *nvr = a->nvram;
+
+       esas2r_trace_enter();
+
+       esas2r_lock_clear_flags(&a->flags, AF_DISC_IN_PROG);
+       esas2r_lock_clear_flags(&a->flags2, AF2_DEV_SCAN);
+       esas2r_lock_clear_flags(&a->flags2, AF2_DEV_CNT_OK);
+
+       a->disc_start_time = jiffies_to_msecs(jiffies);
+       a->disc_wait_time = nvr->dev_wait_time * 1000;
+       a->disc_wait_cnt = nvr->dev_wait_count;
+
+       if (a->disc_wait_cnt > ESAS2R_MAX_TARGETS)
+               a->disc_wait_cnt = ESAS2R_MAX_TARGETS;
+
+       /*
+        * If we are doing chip reset or power management processing, always
+        * wait for devices.  use the NVRAM device count if it is greater than
+        * previously discovered devices.
+        */
+
+       esas2r_hdebug("starting discovery...");
+
+       a->general_req->interrupt_cx = NULL;
+
+       if (a->flags & (AF_CHPRST_DETECTED | AF_POWER_MGT)) {
+               if (a->prev_dev_cnt == 0) {
+                       /* Don't bother waiting if there is nothing to wait
+                        * for.
+                        */
+                       a->disc_wait_time = 0;
+               } else {
+                       /*
+                        * Set the device wait count to what was previously
+                        * found.  We don't care if the user only configured
+                        * a time because we know the exact count to wait for.
+                        * There is no need to honor the user's wishes to
+                        * always wait the full time.
+                        */
+                       a->disc_wait_cnt = a->prev_dev_cnt;
+
+                       /*
+                        * bump the minimum wait time to 15 seconds since the
+                        * default is 3 (system boot or the boot driver usually
+                        * buys us more time).
+                        */
+                       if (a->disc_wait_time < 15000)
+                               a->disc_wait_time = 15000;
+               }
+       }
+
+       esas2r_trace("disc wait count: %d", a->disc_wait_cnt);
+       esas2r_trace("disc wait time: %d", a->disc_wait_time);
+
+       if (a->disc_wait_time == 0)
+               esas2r_disc_check_complete(a);
+
+       esas2r_trace_exit();
+}
+
+void esas2r_disc_start_waiting(struct esas2r_adapter *a)
+{
+       unsigned long flags;
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       if (a->disc_ctx->disc_evt)
+               esas2r_disc_start_port(a);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+}
+
+void esas2r_disc_check_for_work(struct esas2r_adapter *a)
+{
+       struct esas2r_request *rq = a->general_req;
+
+       /* service any pending interrupts first */
+
+       esas2r_polled_interrupt(a);
+
+       /*
+        * now, interrupt processing may have queued up a discovery event.  go
+        * see if we have one to start.  we couldn't start it in the ISR since
+        * polled discovery would cause a deadlock.
+        */
+
+       esas2r_disc_start_waiting(a);
+
+       if (rq->interrupt_cx == NULL)
+               return;
+
+       if (rq->req_stat == RS_STARTED
+           && rq->timeout <= RQ_MAX_TIMEOUT) {
+               /* wait for the current discovery request to complete. */
+               esas2r_wait_request(a, rq);
+
+               if (rq->req_stat == RS_TIMEOUT) {
+                       esas2r_disc_abort(a, rq);
+                       esas2r_local_reset_adapter(a);
+                       return;
+               }
+       }
+
+       if (rq->req_stat == RS_PENDING
+           || rq->req_stat == RS_STARTED)
+               return;
+
+       esas2r_disc_continue(a, rq);
+}
+
+void esas2r_disc_check_complete(struct esas2r_adapter *a)
+{
+       unsigned long flags;
+
+       esas2r_trace_enter();
+
+       /* check to see if we should be waiting for devices */
+       if (a->disc_wait_time) {
+               u32 currtime = jiffies_to_msecs(jiffies);
+               u32 time = currtime - a->disc_start_time;
+
+               /*
+                * Wait until the device wait time is exhausted or the device
+                * wait count is satisfied.
+                */
+               if (time < a->disc_wait_time
+                   && (esas2r_targ_db_get_tgt_cnt(a) < a->disc_wait_cnt
+                       || a->disc_wait_cnt == 0)) {
+                       /* After three seconds of waiting, schedule a scan. */
+                       if (time >= 3000
+                           && !(esas2r_lock_set_flags(&a->flags2,
+                                                      AF2_DEV_SCAN) &
+                                ilog2(AF2_DEV_SCAN))) {
+                               spin_lock_irqsave(&a->mem_lock, flags);
+                               esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
+                               spin_unlock_irqrestore(&a->mem_lock, flags);
+                       }
+
+                       esas2r_trace_exit();
+                       return;
+               }
+
+               /*
+                * We are done waiting...we think.  Adjust the wait time to
+                * consume events after the count is met.
+                */
+               if (!(esas2r_lock_set_flags(&a->flags2, AF2_DEV_CNT_OK)
+                     & ilog2(AF2_DEV_CNT_OK)))
+                       a->disc_wait_time = time + 3000;
+
+               /* If we haven't done a full scan yet, do it now. */
+               if (!(esas2r_lock_set_flags(&a->flags2,
+                                           AF2_DEV_SCAN) &
+                     ilog2(AF2_DEV_SCAN))) {
+                       spin_lock_irqsave(&a->mem_lock, flags);
+                       esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
+                       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+                       esas2r_trace_exit();
+                       return;
+               }
+
+               /*
+                * Now, if there is still time left to consume events, continue
+                * waiting.
+                */
+               if (time < a->disc_wait_time) {
+                       esas2r_trace_exit();
+                       return;
+               }
+       } else {
+               if (!(esas2r_lock_set_flags(&a->flags2,
+                                           AF2_DEV_SCAN) &
+                     ilog2(AF2_DEV_SCAN))) {
+                       spin_lock_irqsave(&a->mem_lock, flags);
+                       esas2r_disc_queue_event(a, DCDE_DEV_SCAN);
+                       spin_unlock_irqrestore(&a->mem_lock, flags);
+               }
+       }
+
+       /* We want to stop waiting for devices. */
+       a->disc_wait_time = 0;
+
+       if ((a->flags & AF_DISC_POLLED)
+           && (a->flags & AF_DISC_IN_PROG)) {
+               /*
+                * Polled discovery is still pending so continue the active
+                * discovery until it is done.  At that point, we will stop
+                * polled discovery and transition to interrupt driven
+                * discovery.
+                */
+       } else {
+               /*
+                * Done waiting for devices.  Note that we get here immediately
+                * after deferred waiting completes because that is interrupt
+                * driven; i.e. There is no transition.
+                */
+               esas2r_disc_fix_curr_requests(a);
+               esas2r_lock_clear_flags(&a->flags, AF_DISC_PENDING);
+
+               /*
+                * We have deferred target state changes until now because we
+                * don't want to report any removals (due to the first arrival)
+                * until the device wait time expires.
+                */
+               esas2r_lock_set_flags(&a->flags, AF_PORT_CHANGE);
+       }
+
+       esas2r_trace_exit();
+}
+
+void esas2r_disc_queue_event(struct esas2r_adapter *a, u8 disc_evt)
+{
+       struct esas2r_disc_context *dc = a->disc_ctx;
+
+       esas2r_trace_enter();
+
+       esas2r_trace("disc_event: %d", disc_evt);
+
+       /* Initialize the discovery context */
+       dc->disc_evt |= disc_evt;
+
+       /*
+        * Don't start discovery before or during polled discovery.  if we did,
+        * we would have a deadlock if we are in the ISR already.
+        */
+       if (!(a->flags & (AF_CHPRST_PENDING | AF_DISC_POLLED)))
+               esas2r_disc_start_port(a);
+
+       esas2r_trace_exit();
+}
+
+bool esas2r_disc_start_port(struct esas2r_adapter *a)
+{
+       struct esas2r_request *rq = a->general_req;
+       struct esas2r_disc_context *dc = a->disc_ctx;
+       bool ret;
+
+       esas2r_trace_enter();
+
+       if (a->flags & AF_DISC_IN_PROG) {
+               esas2r_trace_exit();
+
+               return false;
+       }
+
+       /* If there is a discovery waiting, process it. */
+       if (dc->disc_evt) {
+               if ((a->flags & AF_DISC_POLLED)
+                   && a->disc_wait_time == 0) {
+                       /*
+                        * We are doing polled discovery, but we no longer want
+                        * to wait for devices.  Stop polled discovery and
+                        * transition to interrupt driven discovery.
+                        */
+
+                       esas2r_trace_exit();
+
+                       return false;
+               }
+       } else {
+               /* Discovery is complete. */
+
+               esas2r_hdebug("disc done");
+
+               esas2r_lock_set_flags(&a->flags, AF_PORT_CHANGE);
+
+               esas2r_trace_exit();
+
+               return false;
+       }
+
+       /* Handle the discovery context */
+       esas2r_trace("disc_evt: %d", dc->disc_evt);
+       esas2r_lock_set_flags(&a->flags, AF_DISC_IN_PROG);
+       dc->flags = 0;
+
+       if (a->flags & AF_DISC_POLLED)
+               dc->flags |= DCF_POLLED;
+
+       rq->interrupt_cx = dc;
+       rq->req_stat = RS_SUCCESS;
+
+       /* Decode the event code */
+       if (dc->disc_evt & DCDE_DEV_SCAN) {
+               dc->disc_evt &= ~DCDE_DEV_SCAN;
+
+               dc->flags |= DCF_DEV_SCAN;
+               dc->state = DCS_BLOCK_DEV_SCAN;
+       } else if (dc->disc_evt & DCDE_DEV_CHANGE) {
+               dc->disc_evt &= ~DCDE_DEV_CHANGE;
+
+               dc->flags |= DCF_DEV_CHANGE;
+               dc->state = DCS_DEV_RMV;
+       }
+
+       /* Continue interrupt driven discovery */
+       if (!(a->flags & AF_DISC_POLLED))
+               ret = esas2r_disc_continue(a, rq);
+       else
+               ret = true;
+
+       esas2r_trace_exit();
+
+       return ret;
+}
+
+static bool esas2r_disc_continue(struct esas2r_adapter *a,
+                                struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+
+       /* Device discovery/removal */
+       while (dc->flags & (DCF_DEV_CHANGE | DCF_DEV_SCAN)) {
+               rslt = false;
+
+               switch (dc->state) {
+               case DCS_DEV_RMV:
+
+                       rslt = esas2r_disc_dev_remove(a, rq);
+                       break;
+
+               case DCS_DEV_ADD:
+
+                       rslt = esas2r_disc_dev_add(a, rq);
+                       break;
+
+               case DCS_BLOCK_DEV_SCAN:
+
+                       rslt = esas2r_disc_block_dev_scan(a, rq);
+                       break;
+
+               case DCS_RAID_GRP_INFO:
+
+                       rslt = esas2r_disc_raid_grp_info(a, rq);
+                       break;
+
+               case DCS_PART_INFO:
+
+                       rslt = esas2r_disc_part_info(a, rq);
+                       break;
+
+               case DCS_PT_DEV_INFO:
+
+                       rslt = esas2r_disc_passthru_dev_info(a, rq);
+                       break;
+               case DCS_PT_DEV_ADDR:
+
+                       rslt = esas2r_disc_passthru_dev_addr(a, rq);
+                       break;
+               case DCS_DISC_DONE:
+
+                       dc->flags &= ~(DCF_DEV_CHANGE | DCF_DEV_SCAN);
+                       break;
+
+               default:
+
+                       esas2r_bugon();
+                       dc->state = DCS_DISC_DONE;
+                       break;
+               }
+
+               if (rslt)
+                       return true;
+       }
+
+       /* Discovery is done...for now. */
+       rq->interrupt_cx = NULL;
+
+       if (!(a->flags & AF_DISC_PENDING))
+               esas2r_disc_fix_curr_requests(a);
+
+       esas2r_lock_clear_flags(&a->flags, AF_DISC_IN_PROG);
+
+       /* Start the next discovery. */
+       return esas2r_disc_start_port(a);
+}
+
+static bool esas2r_disc_start_request(struct esas2r_adapter *a,
+                                     struct esas2r_request *rq)
+{
+       unsigned long flags;
+
+       /* Set the timeout to a minimum value. */
+       if (rq->timeout < ESAS2R_DEFAULT_TMO)
+               rq->timeout = ESAS2R_DEFAULT_TMO;
+
+       /*
+        * Override the request type to distinguish discovery requests.  If we
+        * end up deferring the request, esas2r_disc_local_start_request()
+        * will be called to restart it.
+        */
+       rq->req_type = RT_DISC_REQ;
+
+       spin_lock_irqsave(&a->queue_lock, flags);
+
+       if (!(a->flags & (AF_CHPRST_PENDING | AF_FLASHING)))
+               esas2r_disc_local_start_request(a, rq);
+       else
+               esas2r_enqueue(a->dummy_defer, rq);
+
+       spin_unlock_irqrestore(&a->queue_lock, flags);
+
+       return true;
+}
+
+void esas2r_disc_local_start_request(struct esas2r_adapter *a,
+                                    struct esas2r_request *rq)
+{
+       esas2r_trace_enter();
+
+       esas2r_enqueue(a->dummy_active, rq);
+
+       esas2r_start_vda_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return;
+}
+
+static void esas2r_disc_abort(struct esas2r_adapter *a,
+                             struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+
+       esas2r_trace_enter();
+
+       /* abort the current discovery */
+
+       dc->state = DCS_DISC_DONE;
+
+       esas2r_trace_exit();
+}
+
+static bool esas2r_disc_block_dev_scan(struct esas2r_adapter *a,
+                                      struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+
+       esas2r_trace_enter();
+
+       esas2r_rq_init_request(rq, a);
+
+       esas2r_build_mgt_req(a,
+                            rq,
+                            VDAMGT_DEV_SCAN,
+                            0,
+                            0,
+                            0,
+                            NULL);
+
+       rq->comp_cb = esas2r_disc_block_dev_scan_cb;
+
+       rq->timeout = 30000;
+       rq->interrupt_cx = dc;
+
+       rslt = esas2r_disc_start_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return rslt;
+}
+
+static void esas2r_disc_block_dev_scan_cb(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       unsigned long flags;
+
+       esas2r_trace_enter();
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       if (rq->req_stat == RS_SUCCESS)
+               dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
+
+       dc->state = DCS_RAID_GRP_INFO;
+       dc->raid_grp_ix = 0;
+
+       esas2r_rq_destroy_request(rq, a);
+
+       /* continue discovery if it's interrupt driven */
+
+       if (!(dc->flags & DCF_POLLED))
+               esas2r_disc_continue(a, rq);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+       esas2r_trace_exit();
+}
+
+static bool esas2r_disc_raid_grp_info(struct esas2r_adapter *a,
+                                     struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+       struct atto_vda_grp_info *grpinfo;
+
+       esas2r_trace_enter();
+
+       esas2r_trace("raid_group_idx: %d", dc->raid_grp_ix);
+
+       if (dc->raid_grp_ix >= VDA_MAX_RAID_GROUPS) {
+               dc->state = DCS_DISC_DONE;
+
+               esas2r_trace_exit();
+
+               return false;
+       }
+
+       esas2r_rq_init_request(rq, a);
+
+       grpinfo = &rq->vda_rsp_data->mgt_data.data.grp_info;
+
+       memset(grpinfo, 0, sizeof(struct atto_vda_grp_info));
+
+       esas2r_build_mgt_req(a,
+                            rq,
+                            VDAMGT_GRP_INFO,
+                            dc->scan_gen,
+                            0,
+                            sizeof(struct atto_vda_grp_info),
+                            NULL);
+
+       grpinfo->grp_index = dc->raid_grp_ix;
+
+       rq->comp_cb = esas2r_disc_raid_grp_info_cb;
+
+       rq->interrupt_cx = dc;
+
+       rslt = esas2r_disc_start_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return rslt;
+}
+
+static void esas2r_disc_raid_grp_info_cb(struct esas2r_adapter *a,
+                                        struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       unsigned long flags;
+       struct atto_vda_grp_info *grpinfo;
+
+       esas2r_trace_enter();
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       if (rq->req_stat == RS_SCAN_GEN) {
+               dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
+               dc->raid_grp_ix = 0;
+               goto done;
+       }
+
+       if (rq->req_stat == RS_SUCCESS) {
+               grpinfo = &rq->vda_rsp_data->mgt_data.data.grp_info;
+
+               if (grpinfo->status != VDA_GRP_STAT_ONLINE
+                   && grpinfo->status != VDA_GRP_STAT_DEGRADED) {
+                       /* go to the next group. */
+
+                       dc->raid_grp_ix++;
+               } else {
+                       memmove(&dc->raid_grp_name[0],
+                               &grpinfo->grp_name[0],
+                               sizeof(grpinfo->grp_name));
+
+                       dc->interleave = le32_to_cpu(grpinfo->interleave);
+                       dc->block_size = le32_to_cpu(grpinfo->block_size);
+
+                       dc->state = DCS_PART_INFO;
+                       dc->part_num = 0;
+               }
+       } else {
+               if (!(rq->req_stat == RS_GRP_INVALID)) {
+                       esas2r_log(ESAS2R_LOG_WARN,
+                                  "A request for RAID group info failed - "
+                                  "returned with %x",
+                                  rq->req_stat);
+               }
+
+               dc->dev_ix = 0;
+               dc->state = DCS_PT_DEV_INFO;
+       }
+
+done:
+
+       esas2r_rq_destroy_request(rq, a);
+
+       /* continue discovery if it's interrupt driven */
+
+       if (!(dc->flags & DCF_POLLED))
+               esas2r_disc_continue(a, rq);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+       esas2r_trace_exit();
+}
+
+static bool esas2r_disc_part_info(struct esas2r_adapter *a,
+                                 struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+       struct atto_vdapart_info *partinfo;
+
+       esas2r_trace_enter();
+
+       esas2r_trace("part_num: %d", dc->part_num);
+
+       if (dc->part_num >= VDA_MAX_PARTITIONS) {
+               dc->state = DCS_RAID_GRP_INFO;
+               dc->raid_grp_ix++;
+
+               esas2r_trace_exit();
+
+               return false;
+       }
+
+       esas2r_rq_init_request(rq, a);
+
+       partinfo = &rq->vda_rsp_data->mgt_data.data.part_info;
+
+       memset(partinfo, 0, sizeof(struct atto_vdapart_info));
+
+       esas2r_build_mgt_req(a,
+                            rq,
+                            VDAMGT_PART_INFO,
+                            dc->scan_gen,
+                            0,
+                            sizeof(struct atto_vdapart_info),
+                            NULL);
+
+       partinfo->part_no = dc->part_num;
+
+       memmove(&partinfo->grp_name[0],
+               &dc->raid_grp_name[0],
+               sizeof(partinfo->grp_name));
+
+       rq->comp_cb = esas2r_disc_part_info_cb;
+
+       rq->interrupt_cx = dc;
+
+       rslt = esas2r_disc_start_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return rslt;
+}
+
+static void esas2r_disc_part_info_cb(struct esas2r_adapter *a,
+                                    struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       unsigned long flags;
+       struct atto_vdapart_info *partinfo;
+
+       esas2r_trace_enter();
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       if (rq->req_stat == RS_SCAN_GEN) {
+               dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
+               dc->raid_grp_ix = 0;
+               dc->state = DCS_RAID_GRP_INFO;
+       } else if (rq->req_stat == RS_SUCCESS) {
+               partinfo = &rq->vda_rsp_data->mgt_data.data.part_info;
+
+               dc->part_num = partinfo->part_no;
+
+               dc->curr_virt_id = le16_to_cpu(partinfo->target_id);
+
+               esas2r_targ_db_add_raid(a, dc);
+
+               dc->part_num++;
+       } else {
+               if (!(rq->req_stat == RS_PART_LAST)) {
+                       esas2r_log(ESAS2R_LOG_WARN,
+                                  "A request for RAID group partition info "
+                                  "failed - status:%d", rq->req_stat);
+               }
+
+               dc->state = DCS_RAID_GRP_INFO;
+               dc->raid_grp_ix++;
+       }
+
+       esas2r_rq_destroy_request(rq, a);
+
+       /* continue discovery if it's interrupt driven */
+
+       if (!(dc->flags & DCF_POLLED))
+               esas2r_disc_continue(a, rq);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+       esas2r_trace_exit();
+}
+
+static bool esas2r_disc_passthru_dev_info(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+       struct atto_vda_devinfo *devinfo;
+
+       esas2r_trace_enter();
+
+       esas2r_trace("dev_ix: %d", dc->dev_ix);
+
+       esas2r_rq_init_request(rq, a);
+
+       devinfo = &rq->vda_rsp_data->mgt_data.data.dev_info;
+
+       memset(devinfo, 0, sizeof(struct atto_vda_devinfo));
+
+       esas2r_build_mgt_req(a,
+                            rq,
+                            VDAMGT_DEV_PT_INFO,
+                            dc->scan_gen,
+                            dc->dev_ix,
+                            sizeof(struct atto_vda_devinfo),
+                            NULL);
+
+       rq->comp_cb = esas2r_disc_passthru_dev_info_cb;
+
+       rq->interrupt_cx = dc;
+
+       rslt = esas2r_disc_start_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return rslt;
+}
+
+static void esas2r_disc_passthru_dev_info_cb(struct esas2r_adapter *a,
+                                            struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       unsigned long flags;
+       struct atto_vda_devinfo *devinfo;
+
+       esas2r_trace_enter();
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       if (rq->req_stat == RS_SCAN_GEN) {
+               dc->scan_gen = rq->func_rsp.mgt_rsp.scan_generation;
+               dc->dev_ix = 0;
+               dc->state = DCS_PT_DEV_INFO;
+       } else if (rq->req_stat == RS_SUCCESS) {
+               devinfo = &rq->vda_rsp_data->mgt_data.data.dev_info;
+
+               dc->dev_ix = le16_to_cpu(rq->func_rsp.mgt_rsp.dev_index);
+
+               dc->curr_virt_id = le16_to_cpu(devinfo->target_id);
+
+               if (le16_to_cpu(devinfo->features) & VDADEVFEAT_PHYS_ID) {
+                       dc->curr_phys_id =
+                               le16_to_cpu(devinfo->phys_target_id);
+                       dc->dev_addr_type = ATTO_GDA_AT_PORT;
+                       dc->state = DCS_PT_DEV_ADDR;
+
+                       esas2r_trace("curr_virt_id: %d", dc->curr_virt_id);
+                       esas2r_trace("curr_phys_id: %d", dc->curr_phys_id);
+               } else {
+                       dc->dev_ix++;
+               }
+       } else {
+               if (!(rq->req_stat == RS_DEV_INVALID)) {
+                       esas2r_log(ESAS2R_LOG_WARN,
+                                  "A request for device information failed - "
+                                  "status:%d", rq->req_stat);
+               }
+
+               dc->state = DCS_DISC_DONE;
+       }
+
+       esas2r_rq_destroy_request(rq, a);
+
+       /* continue discovery if it's interrupt driven */
+
+       if (!(dc->flags & DCF_POLLED))
+               esas2r_disc_continue(a, rq);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+       esas2r_trace_exit();
+}
+
+static bool esas2r_disc_passthru_dev_addr(struct esas2r_adapter *a,
+                                         struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       bool rslt;
+       struct atto_ioctl *hi;
+       struct esas2r_sg_context sgc;
+
+       esas2r_trace_enter();
+
+       esas2r_rq_init_request(rq, a);
+
+       /* format the request. */
+
+       sgc.cur_offset = NULL;
+       sgc.get_phys_addr = (PGETPHYSADDR)esas2r_disc_get_phys_addr;
+       sgc.length = offsetof(struct atto_ioctl, data)
+                    + sizeof(struct atto_hba_get_device_address);
+
+       esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge);
+
+       esas2r_build_ioctl_req(a, rq, sgc.length, VDA_IOCTL_HBA);
+
+       if (!esas2r_build_sg_list(a, rq, &sgc)) {
+               esas2r_rq_destroy_request(rq, a);
+
+               esas2r_trace_exit();
+
+               return false;
+       }
+
+       rq->comp_cb = esas2r_disc_passthru_dev_addr_cb;
+
+       rq->interrupt_cx = dc;
+
+       /* format the IOCTL data. */
+
+       hi = (struct atto_ioctl *)a->disc_buffer;
+
+       memset(a->disc_buffer, 0, ESAS2R_DISC_BUF_LEN);
+
+       hi->version = ATTO_VER_GET_DEV_ADDR0;
+       hi->function = ATTO_FUNC_GET_DEV_ADDR;
+       hi->flags = HBAF_TUNNEL;
+
+       hi->data.get_dev_addr.target_id = le32_to_cpu(dc->curr_phys_id);
+       hi->data.get_dev_addr.addr_type = dc->dev_addr_type;
+
+       /* start it up. */
+
+       rslt = esas2r_disc_start_request(a, rq);
+
+       esas2r_trace_exit();
+
+       return rslt;
+}
+
+static void esas2r_disc_passthru_dev_addr_cb(struct esas2r_adapter *a,
+                                            struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       struct esas2r_target *t = NULL;
+       unsigned long flags;
+       struct atto_ioctl *hi;
+       u16 addrlen;
+
+       esas2r_trace_enter();
+
+       spin_lock_irqsave(&a->mem_lock, flags);
+
+       hi = (struct atto_ioctl *)a->disc_buffer;
+
+       if (rq->req_stat == RS_SUCCESS
+           && hi->status == ATTO_STS_SUCCESS) {
+               addrlen = le16_to_cpu(hi->data.get_dev_addr.addr_len);
+
+               if (dc->dev_addr_type == ATTO_GDA_AT_PORT) {
+                       if (addrlen == sizeof(u64))
+                               memmove(&dc->sas_addr,
+                                       &hi->data.get_dev_addr.address[0],
+                                       addrlen);
+                       else
+                               memset(&dc->sas_addr, 0, sizeof(dc->sas_addr));
+
+
+                       /* Get the unique identifier. */
+                       dc->dev_addr_type = ATTO_GDA_AT_UNIQUE;
+
+                       goto next_dev_addr;
+               } else {
+                       /* Add the pass through target. */
+                       if (HIBYTE(addrlen) == 0) {
+                               t = esas2r_targ_db_add_pthru(a,
+                                                            dc,
+                                                            &hi->data.
+                                                            get_dev_addr.
+                                                            address[0],
+                                                            (u8)hi->data.
+                                                            get_dev_addr.
+                                                            addr_len);
+
+                               if (t)
+                                       memmove(&t->sas_addr, &dc->sas_addr,
+                                               sizeof(t->sas_addr));
+                       } else {
+                               /* getting the back end data failed */
+
+                               esas2r_log(ESAS2R_LOG_WARN,
+                                          "an error occurred retrieving the "
+                                          "back end data (%s:%d)",
+                                          __func__,
+                                          __LINE__);
+                       }
+               }
+       } else {
+               /* getting the back end data failed */
+
+               esas2r_log(ESAS2R_LOG_WARN,
+                          "an error occurred retrieving the back end data - "
+                          "rq->req_stat:%d hi->status:%d",
+                          rq->req_stat, hi->status);
+       }
+
+       /* proceed to the next device. */
+
+       if (dc->flags & DCF_DEV_SCAN) {
+               dc->dev_ix++;
+               dc->state = DCS_PT_DEV_INFO;
+       } else if (dc->flags & DCF_DEV_CHANGE) {
+               dc->curr_targ++;
+               dc->state = DCS_DEV_ADD;
+       } else {
+               esas2r_bugon();
+       }
+
+next_dev_addr:
+       esas2r_rq_destroy_request(rq, a);
+
+       /* continue discovery if it's interrupt driven */
+
+       if (!(dc->flags & DCF_POLLED))
+               esas2r_disc_continue(a, rq);
+
+       spin_unlock_irqrestore(&a->mem_lock, flags);
+
+       esas2r_trace_exit();
+}
+
+static u32 esas2r_disc_get_phys_addr(struct esas2r_sg_context *sgc, u64 *addr)
+{
+       struct esas2r_adapter *a = sgc->adapter;
+
+       if (sgc->length > ESAS2R_DISC_BUF_LEN)
+               esas2r_bugon();
+
+       *addr = a->esas2r_uncached_phys
+               + (u64)((u8 *)a->disc_buffer - a->esas2r_uncached);
+
+       return sgc->length;
+}
+
+static bool esas2r_disc_dev_remove(struct esas2r_adapter *a,
+                                  struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       struct esas2r_target *t;
+       struct esas2r_target *t2;
+
+       esas2r_trace_enter();
+
+       /* process removals. */
+
+       for (t = a->targetdb; t < a->targetdb_end; t++) {
+               if (t->new_target_state != TS_NOT_PRESENT)
+                       continue;
+
+               t->new_target_state = TS_INVALID;
+
+               /* remove the right target! */
+
+               t2 =
+                       esas2r_targ_db_find_by_virt_id(a,
+                                                      esas2r_targ_get_id(t,
+                                                                         a));
+
+               if (t2)
+                       esas2r_targ_db_remove(a, t2);
+       }
+
+       /* removals complete.  process arrivals. */
+
+       dc->state = DCS_DEV_ADD;
+       dc->curr_targ = a->targetdb;
+
+       esas2r_trace_exit();
+
+       return false;
+}
+
+static bool esas2r_disc_dev_add(struct esas2r_adapter *a,
+                               struct esas2r_request *rq)
+{
+       struct esas2r_disc_context *dc =
+               (struct esas2r_disc_context *)rq->interrupt_cx;
+       struct esas2r_target *t = dc->curr_targ;
+
+       if (t >= a->targetdb_end) {
+               /* done processing state changes. */
+
+               dc->state = DCS_DISC_DONE;
+       } else if (t->new_target_state == TS_PRESENT) {
+               struct atto_vda_ae_lu *luevt = &t->lu_event;
+
+               esas2r_trace_enter();
+
+               /* clear this now in case more events come in. */
+
+               t->new_target_state = TS_INVALID;
+
+               /* setup the discovery context for adding this device. */
+
+               dc->curr_virt_id = esas2r_targ_get_id(t, a);
+
+               if ((luevt->hdr.bylength >= offsetof(struct atto_vda_ae_lu, id)
+                    + sizeof(struct atto_vda_ae_lu_tgt_lun_raid))
+                   && !(luevt->dwevent & VDAAE_LU_PASSTHROUGH)) {
+                       dc->block_size = luevt->id.tgtlun_raid.dwblock_size;
+                       dc->interleave = luevt->id.tgtlun_raid.dwinterleave;
+               } else {
+                       dc->block_size = 0;
+                       dc->interleave = 0;
+               }
+
+               /* determine the device type being added. */
+
+               if (luevt->dwevent & VDAAE_LU_PASSTHROUGH) {
+                       if (luevt->dwevent & VDAAE_LU_PHYS_ID) {
+                               dc->state = DCS_PT_DEV_ADDR;
+                               dc->dev_addr_type = ATTO_GDA_AT_PORT;
+                               dc->curr_phys_id = luevt->wphys_target_id;
+                       } else {
+                               esas2r_log(ESAS2R_LOG_WARN,
+                                          "luevt->dwevent does not have the "
+                                          "VDAAE_LU_PHYS_ID bit set (%s:%d)",
+                                          __func__, __LINE__);
+                       }
+               } else {
+                       dc->raid_grp_name[0] = 0;
+
+                       esas2r_targ_db_add_raid(a, dc);
+               }
+
+               esas2r_trace("curr_virt_id: %d", dc->curr_virt_id);
+               esas2r_trace("curr_phys_id: %d", dc->curr_phys_id);
+               esas2r_trace("dwevent: %d", luevt->dwevent);
+
+               esas2r_trace_exit();
+       }
+
+       if (dc->state == DCS_DEV_ADD) {
+               /* go to the next device. */
+
+               dc->curr_targ++;
+       }
+
+       return false;
+}
+
+/*
+ * When discovery is done, find all requests on defer queue and
+ * test if they need to be modified. If a target is no longer present
+ * then complete the request with RS_SEL. Otherwise, update the
+ * target_id since after a hibernate it can be a different value.
+ * VDA does not make passthrough target IDs persistent.
+ */
+static void esas2r_disc_fix_curr_requests(struct esas2r_adapter *a)
+{
+       unsigned long flags;
+       struct esas2r_target *t;
+       struct esas2r_request *rq;
+
+       /* update virt_targ_id in any outstanding esas2r_requests  */
+
+       spin_lock_irqsave(&a->queue_lock, flags);
+
+       rq = a->first_defer;
+
+       while (rq != a->dummy_defer) {
+               if (rq->vrq->scsi.function == VDA_FUNC_SCSI) {
+                       t = a->targetdb + rq->target_id;
+
+                       if (t->target_state == TS_PRESENT)
+                               rq->vrq->scsi.target_id = le16_to_cpu(
+                                       t->virt_targ_id);
+                       else
+                               rq->req_stat = RS_SEL;
+               }
+
+               rq = rq->next;
+       }
+
+       spin_unlock_irqrestore(&a->queue_lock, flags);
+}
diff --git a/drivers/scsi/esas2r/esas2r_log.c b/drivers/scsi/esas2r/esas2r_log.c
new file mode 100644
index 0000000..651f52c
--- /dev/null
+++ b/drivers/scsi/esas2r/esas2r_log.c
@@ -0,0 +1,254 @@
+/*
+ *  linux/drivers/scsi/esas2r/esas2r_log.c
+ *      For use with ATTO ExpressSAS R6xx SAS/SATA RAID controllers
+ *
+ *  Copyright (c) 2001-2013 ATTO Technology, Inc.
+ *  (mailto:linuxdriv...@attotech.com)
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * NO WARRANTY
+ * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
+ * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
+ * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
+ * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
+ * solely responsible for determining the appropriateness of using and
+ * distributing the Program and assumes all risks associated with its
+ * exercise of rights under this Agreement, including but not limited to
+ * the risks and costs of program errors, damage to or loss of data,
+ * programs or equipment, and unavailability or interruption of operations.
+ *
+ * DISCLAIMER OF LIABILITY
+ * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
+ * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
+ * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
+ * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
+ * USA.
+ */
+
+#include "esas2r.h"
+
+/*
+ * this module within the driver is tasked with providing logging 
functionality.
+ * the event_log_level module parameter controls the level of messages that are
+ * written to the system log.  the default level of messages that are written
+ * are critical and warning messages.  if other types of messages are desired,
+ * one simply needs to load the module with the correct value for the
+ * event_log_level module parameter.  for example:
+ *
+ * insmod <module> event_log_level=1
+ *
+ * will load the module and only critical events will be written by this module
+ * to the system log.  if critical, warning, and information-level messages are
+ * desired, the correct value for the event_log_level module parameter
+ * would be as follows:
+ *
+ * insmod <module> event_log_level=3
+ */
+
+#define EVENT_LOG_BUFF_SIZE 1024
+
+static long event_log_level = ESAS2R_LOG_DFLT;
+
+module_param(event_log_level, long, S_IRUGO | S_IRUSR);
+MODULE_PARM_DESC(event_log_level,
+                "Specifies the level of events to report to the system log.  
Critical and warning level events are logged by default.");
+
+/* A shared buffer to use for formatting messages. */
+static char event_buffer[EVENT_LOG_BUFF_SIZE];
+
+/* A lock to protect the shared buffer used for formatting messages. */
+static DEFINE_SPINLOCK(event_buffer_lock);
+
+/**
+ * translates an esas2r-defined logging event level to a kernel logging level.
+ *
+ * @param [in] level the esas2r-defined logging event level to translate
+ *
+ * @return the corresponding kernel logging level.
+ */
+static const char *translate_esas2r_event_level_to_kernel(const long level)
+{
+       switch (level) {
+       case ESAS2R_LOG_CRIT:
+               return KERN_CRIT;
+
+       case ESAS2R_LOG_WARN:
+               return KERN_WARNING;
+
+       case ESAS2R_LOG_INFO:
+               return KERN_INFO;
+
+       case ESAS2R_LOG_DEBG:
+       case ESAS2R_LOG_TRCE:
+       default:
+               return KERN_DEBUG;
+       }
+}
+
+/**
+ * the master logging function.  this function will format the message as
+ * outlined by the formatting string, the input device information and the
+ * substitution arguments and output the resulting string to the system log.
+ *
+ * @param [in] level  the event log level of the message
+ * @param [in] dev    the device information
+ * @param [in] format the formatting string for the message
+ * @param [in] args   the substition arguments to the formatting string
+ *
+ * @return 0 on success, or -1 if an error occurred.
+ */
+static int esas2r_log_master(const long level,
+                            const struct device *dev,
+                            const char *format,
+                            va_list args)
+{
+       if (level <= event_log_level) {
+               unsigned long flags = 0;
+               int retval = 0;
+               char *buffer = event_buffer;
+               size_t buflen = EVENT_LOG_BUFF_SIZE;
+               const char *fmt_nodev = "%s%s: ";
+               const char *fmt_dev = "%s%s [%s, %s, %s]";
+               const char *slevel =
+                       translate_esas2r_event_level_to_kernel(level);
+
+               spin_lock_irqsave(&event_buffer_lock, flags);
+
+               if (buffer == NULL) {
+                       spin_unlock_irqrestore(&event_buffer_lock, flags);
+                       return -1;
+               }
+
+               memset(buffer, 0, buflen);
+
+               /*
+                * format the level onto the beginning of the string and do
+                * some pointer arithmetic to move the pointer to the point
+                * where the actual message can be inserted.
+                */
+
+               if (dev == NULL) {
+                       snprintf(buffer, buflen, fmt_nodev, slevel,
+                                ESAS2R_DRVR_NAME);
+               } else {
+                       snprintf(buffer, buflen, fmt_dev, slevel,
+                                ESAS2R_DRVR_NAME,
+                                (dev->driver ? dev->driver->name : "unknown"),
+                                (dev->bus ? dev->bus->name : "unknown"),
+                                dev_name(dev));
+               }
+
+               buffer += strlen(event_buffer);
+               buflen -= strlen(event_buffer);
+
+               retval = vsnprintf(buffer, buflen, format, args);
+               if (retval < 0) {
+                       spin_unlock_irqrestore(&event_buffer_lock, flags);
+                       return -1;
+               }
+
+               /*
+                * Put a line break at the end of the formatted string so that
+                * we don't wind up with run-on messages.  only append if there
+                * is enough space in the buffer.
+                */
+               if (strlen(event_buffer) < buflen)
+                       strcat(buffer, "\n");
+
+               printk(event_buffer);
+
+               spin_unlock_irqrestore(&event_buffer_lock, flags);
+       }
+
+       return 0;
+}
+
+/**
+ * formats and logs a message to the system log.
+ *
+ * @param [in] level  the event level of the message
+ * @param [in] format the formating string for the message
+ * @param [in] ...    the substitution arguments to the formatting string
+ *
+ * @return 0 on success, or -1 if an error occurred.
+ */
+int esas2r_log(const long level, const char *format, ...)
+{
+       int retval = 0;
+       va_list args;
+
+       va_start(args, format);
+
+       retval = esas2r_log_master(level, NULL, format, args);
+
+       va_end(args);
+
+       return retval;
+}
+
+/**
+ * formats and logs a message to the system log.  this message will include
+ * device information.
+ *
+ * @param [in] level   the event level of the message
+ * @param [in] dev     the device information
+ * @param [in] format  the formatting string for the message
+ * @param [in] ...     the substitution arguments to the formatting string
+ *
+ * @return 0 on success, or -1 if an error occurred.
+ */
+int esas2r_log_dev(const long level,
+                  const struct device *dev,
+                  const char *format,
+                  ...)
+{
+       int retval = 0;
+       va_list args;
+
+       va_start(args, format);
+
+       retval = esas2r_log_master(level, dev, format, args);
+
+       va_end(args);
+
+       return retval;
+}
+
+/**
+ * formats and logs a message to the system log.  this message will include
+ * device information.
+ *
+ * @param [in] level   the event level of the message
+ * @param [in] buf
+ * @param [in] len
+ *
+ * @return 0 on success, or -1 if an error occurred.
+ */
+int esas2r_log_hexdump(const long level,
+                      const void *buf,
+                      size_t len)
+{
+       if (level <= event_log_level) {
+               print_hex_dump(translate_esas2r_event_level_to_kernel(level),
+                              "", DUMP_PREFIX_OFFSET, 16, 1, buf,
+                              len, true);
+       }
+
+       return 1;
+}
diff --git a/drivers/scsi/esas2r/esas2r_log.h b/drivers/scsi/esas2r/esas2r_log.h
new file mode 100644
index 0000000..b65770b
--- /dev/null
+++ b/drivers/scsi/esas2r/esas2r_log.h
@@ -0,0 +1,118 @@
+/*
+ *  linux/drivers/scsi/esas2r/esas2r_log.h
+ *      For use with ATTO ExpressSAS R6xx SAS/SATA RAID controllers
+ *
+ *  Copyright (c) 2001-2013 ATTO Technology, Inc.
+ *  (mailto:linuxdriv...@attotech.com)
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * NO WARRANTY
+ * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
+ * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
+ * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
+ * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
+ * solely responsible for determining the appropriateness of using and
+ * distributing the Program and assumes all risks associated with its
+ * exercise of rights under this Agreement, including but not limited to
+ * the risks and costs of program errors, damage to or loss of data,
+ * programs or equipment, and unavailability or interruption of operations.
+ *
+ * DISCLAIMER OF LIABILITY
+ * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
+ * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
+ * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
+ * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
+ * USA.
+ */
+
+#ifndef __esas2r_log_h__
+#define __esas2r_log_h__
+
+struct device;
+
+enum {
+       ESAS2R_LOG_NONE = 0,    /* no events logged */
+       ESAS2R_LOG_CRIT = 1,    /* critical events  */
+       ESAS2R_LOG_WARN = 2,    /* warning events   */
+       ESAS2R_LOG_INFO = 3,    /* info events      */
+       ESAS2R_LOG_DEBG = 4,    /* debugging events */
+       ESAS2R_LOG_TRCE = 5,    /* tracing events   */
+
+#ifdef ESAS2R_TRACE
+       ESAS2R_LOG_DFLT = ESAS2R_LOG_TRCE
+#else
+       ESAS2R_LOG_DFLT = ESAS2R_LOG_WARN
+#endif
+};
+
+int esas2r_log(const long level, const char *format, ...);
+int esas2r_log_dev(const long level,
+                  const struct device *dev,
+                  const char *format,
+                  ...);
+int esas2r_log_hexdump(const long level,
+                      const void *buf,
+                      size_t len);
+
+/*
+ * the following macros are provided specifically for debugging and tracing
+ * messages.  esas2r_debug() is provided for generic non-hardware layer
+ * debugging and tracing events.  esas2r_hdebug is provided specifically for
+ * hardware layer debugging and tracing events.
+ */
+
+#ifdef ESAS2R_DEBUG
+#define esas2r_debug(f, args ...) esas2r_log(ESAS2R_LOG_DEBG, f, ## args)
+#define esas2r_hdebug(f, args ...) esas2r_log(ESAS2R_LOG_DEBG, f, ## args)
+#else
+#define esas2r_debug(f, args ...)
+#define esas2r_hdebug(f, args ...)
+#endif  /* ESAS2R_DEBUG */
+
+/*
+ * the following macros are provided in order to trace the driver and catch
+ * some more serious bugs.  be warned, enabling these macros may *severely*
+ * impact performance.
+ */
+
+#ifdef ESAS2R_TRACE
+#define esas2r_bugon() \
+       do { \
+               esas2r_log(ESAS2R_LOG_TRCE, "esas2r_bugon() called in %s:%d - " 
\
+                          "dumping stack and stopping kernel", __func__, \
+                          __LINE__); \
+               dump_stack(); \
+               BUG(); \
+       } while (0)
+
+#define esas2r_trace_enter() esas2r_log(ESAS2R_LOG_TRCE, "entered %s (%s:%d)", 
\
+                                       __func__, __FILE__, __LINE__)
+#define esas2r_trace_exit() esas2r_log(ESAS2R_LOG_TRCE, "exited %s (%s:%d)", \
+                                      __func__, __FILE__, __LINE__)
+#define esas2r_trace(f, args ...) esas2r_log(ESAS2R_LOG_TRCE, "(%s:%s:%d): " 
f, \
+                                            __func__, __FILE__, __LINE__, \
+                                            ## args)
+#else
+#define esas2r_bugon()
+#define esas2r_trace_enter()
+#define esas2r_trace_exit()
+#define esas2r_trace(f, args ...)
+#endif  /* ESAS2R_TRACE */
+
+#endif  /* __esas2r_log_h__ */
-- 
1.8.1.4

--
To unsubscribe from this list: send the line "unsubscribe linux-scsi" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to