]> bbs.cooldavid.org Git - net-next-2.6.git/commitdiff
Moorestown "Post Beta" nand driver.
authorAlan Olsen <alan.r.olsen@intel.com>
Mon, 26 Apr 2010 17:50:19 +0000 (10:50 -0700)
committerDavid Woodhouse <David.Woodhouse@intel.com>
Mon, 7 Jun 2010 15:23:54 +0000 (16:23 +0100)
Contains the fixes for nand corruption with the watchdog driver.

New features compares to MRST NAND driver Post Alpah2 2.0:
1. Enable CDMA feature of NAND controller

How to use this driver:
The same with before. That is, to enable this driver,
you can set
CONFIG_MRST_NAND=y
CONFIG_MRST_NAND_HW=y

Signed-off-by: Gao Yunpeng <yunpeng.gao@intel.com>
Signed-off-by: Alan Olsen <alan.r.olsen@intel.com>
20 files changed:
drivers/block/Kconfig
drivers/block/Makefile
drivers/block/spectra/Kconfig [new file with mode: 0644]
drivers/block/spectra/Makefile [new file with mode: 0644]
drivers/block/spectra/README [new file with mode: 0644]
drivers/block/spectra/ffsdefs.h [new file with mode: 0644]
drivers/block/spectra/ffsport.c [new file with mode: 0644]
drivers/block/spectra/ffsport.h [new file with mode: 0644]
drivers/block/spectra/flash.c [new file with mode: 0644]
drivers/block/spectra/flash.h [new file with mode: 0644]
drivers/block/spectra/lld.c [new file with mode: 0644]
drivers/block/spectra/lld.h [new file with mode: 0644]
drivers/block/spectra/lld_cdma.c [new file with mode: 0644]
drivers/block/spectra/lld_cdma.h [new file with mode: 0644]
drivers/block/spectra/lld_emu.c [new file with mode: 0644]
drivers/block/spectra/lld_emu.h [new file with mode: 0644]
drivers/block/spectra/lld_nand.c [new file with mode: 0644]
drivers/block/spectra/lld_nand.h [new file with mode: 0644]
drivers/block/spectra/nand_regs.h [new file with mode: 0644]
drivers/block/spectra/spectraswconfig.h [new file with mode: 0644]

index 77bfce52e9cabc04a51d04ed6a9c784969e114ca..d62b95d2ab0055d53b5fa241dc8f43af69f5edba 100644 (file)
@@ -488,4 +488,6 @@ config BLK_DEV_HD
 
          If unsure, say N.
 
+source "drivers/block/spectra/Kconfig"
+
 endif # BLK_DEV
index aff5ac925c34332f235e523164f7f85cd76b6751..568ba651cb5ebc36d18e30d06f69fa810b9a64ee 100644 (file)
@@ -38,4 +38,6 @@ obj-$(CONFIG_BLK_DEV_HD)      += hd.o
 obj-$(CONFIG_XEN_BLKDEV_FRONTEND)      += xen-blkfront.o
 obj-$(CONFIG_BLK_DEV_DRBD)     += drbd/
 
+obj-$(CONFIG_MRST_NAND)        += spectra/
+
 swim_mod-objs  := swim.o swim_asm.o
diff --git a/drivers/block/spectra/Kconfig b/drivers/block/spectra/Kconfig
new file mode 100644 (file)
index 0000000..fbece10
--- /dev/null
@@ -0,0 +1,27 @@
+
+menuconfig MRST_NAND
+       tristate "Moorestown NAND Flash controller"
+       depends on BLOCK
+       default n
+       ---help---
+         Enable the driver for the NAND Flash controller in Intel Moorestown
+         Platform
+
+choice
+       prompt "Compile for"
+       depends on MRST_NAND
+       default MRST_NAND_HW
+
+config MRST_NAND_HW
+       bool "Actual hardware mode"
+       help
+         Driver communicates with the actual hardware's register interface.
+         in DMA mode.
+
+config MRST_NAND_EMU
+       bool "RAM emulator testing"
+       help
+         Driver emulates Flash on a RAM buffer and / or disk file.  Useful to test the behavior of FTL layer.
+
+endchoice
+
diff --git a/drivers/block/spectra/Makefile b/drivers/block/spectra/Makefile
new file mode 100644 (file)
index 0000000..261891c
--- /dev/null
@@ -0,0 +1,7 @@
+#
+# Makefile of Intel Moorestown NAND controller driver
+#
+
+obj-$(CONFIG_MRST_NAND) += spectra.o
+spectra-objs := ffsport.o flash.o lld.o lld_emu.o lld_nand.o lld_cdma.o
+
diff --git a/drivers/block/spectra/README b/drivers/block/spectra/README
new file mode 100644 (file)
index 0000000..ecba559
--- /dev/null
@@ -0,0 +1,29 @@
+This is a driver for NAND controller of Intel Moorestown platform.
+
+This driver is a standalone linux block device driver, it acts as if it's a normal hard disk.
+It includes three layer:
+       block layer interface - file ffsport.c
+       Flash Translation Layer (FTL) - file flash.c (implement the NAND flash Translation Layer, includs address mapping, garbage collection, wear-leveling and so on)
+       Low level layer - file lld_nand.c/lld_cdma.c/lld_emu.c (which implements actual controller hardware registers access)
+
+This driver can be build as modules or build-in.
+
+Dependency:
+This driver has dependency on IA Firmware of Intel Moorestown platform.
+It need the IA Firmware to create the block table for the first time.
+And to validate this driver code without IA Firmware, you can change the
+macro AUTO_FORMAT_FLASH from 0 to 1 in file spectraswconfig.h. Thus the
+driver will erase the whole nand flash and create a new block table.
+
+TODO:
+       - Enable Command DMA feature support
+       - lower the memory footprint
+       - Remove most of the unnecessary global variables
+       - Change all the upcase variable / functions name to lowercase
+       - Some other misc bugs
+
+Please send patches to:
+       Greg Kroah-Hartman <gregkh@suse.de>
+
+And Cc to: Gao Yunpeng <yunpeng.gao@intel.com>
+
diff --git a/drivers/block/spectra/ffsdefs.h b/drivers/block/spectra/ffsdefs.h
new file mode 100644 (file)
index 0000000..a9e9cd2
--- /dev/null
@@ -0,0 +1,58 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _FFSDEFS_
+#define _FFSDEFS_
+
+#define CLEAR 0                        /*use this to clear a field instead of "fail"*/
+#define SET   1                        /*use this to set a field instead of "pass"*/
+#define FAIL 1                 /*failed flag*/
+#define PASS 0                 /*success flag*/
+#define ERR -1                 /*error flag*/
+
+#define   ERASE_CMD             10
+#define   WRITE_MAIN_CMD        11
+#define   READ_MAIN_CMD         12
+#define   WRITE_SPARE_CMD       13
+#define   READ_SPARE_CMD        14
+#define   WRITE_MAIN_SPARE_CMD  15
+#define   READ_MAIN_SPARE_CMD   16
+#define   MEMCOPY_CMD           17
+#define   DUMMY_CMD             99
+
+#define     EVENT_PASS                                  0x00
+#define     EVENT_CORRECTABLE_DATA_ERROR_FIXED         0x01
+#define     EVENT_UNCORRECTABLE_DATA_ERROR              0x02
+#define     EVENT_TIME_OUT                              0x03
+#define     EVENT_PROGRAM_FAILURE                       0x04
+#define     EVENT_ERASE_FAILURE                         0x05
+#define     EVENT_MEMCOPY_FAILURE                       0x06
+#define     EVENT_FAIL                                  0x07
+
+#define     EVENT_NONE                                  0x22
+#define     EVENT_DMA_CMD_COMP                          0x77
+#define     EVENT_ECC_TRANSACTION_DONE                  0x88
+#define     EVENT_DMA_CMD_FAIL                          0x99
+
+#define CMD_PASS        0
+#define CMD_FAIL        1
+#define CMD_ABORT       2
+#define CMD_NOT_DONE    3
+
+#endif /* _FFSDEFS_ */
diff --git a/drivers/block/spectra/ffsport.c b/drivers/block/spectra/ffsport.c
new file mode 100644 (file)
index 0000000..0b3d49d
--- /dev/null
@@ -0,0 +1,847 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include "ffsport.h"
+#include "flash.h"
+#include <linux/interrupt.h>
+#include <linux/delay.h>
+#include <linux/blkdev.h>
+#include <linux/wait.h>
+#include <linux/mutex.h>
+#include <linux/kthread.h>
+#include <linux/log2.h>
+#include <linux/init.h>
+
+/**** Helper functions used for Div, Remainder operation on u64 ****/
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_Calc_Used_Bits
+* Inputs:       Power of 2 number
+* Outputs:      Number of Used Bits
+*               0, if the argument is 0
+* Description:  Calculate the number of bits used by a given power of 2 number
+*               Number can be upto 32 bit
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_Calc_Used_Bits(u32 n)
+{
+       int tot_bits = 0;
+
+       if (n >= 1 << 16) {
+               n >>= 16;
+               tot_bits += 16;
+       }
+
+       if (n >= 1 << 8) {
+               n >>=  8;
+               tot_bits +=  8;
+       }
+
+       if (n >= 1 << 4) {
+               n >>=  4;
+               tot_bits +=  4;
+       }
+
+       if (n >= 1 << 2) {
+               n >>=  2;
+               tot_bits +=  2;
+       }
+
+       if (n >= 1 << 1)
+               tot_bits +=  1;
+
+       return ((n == 0) ? (0) : tot_bits);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_u64_Div
+* Inputs:       Number of u64
+*               A power of 2 number as Division
+* Outputs:      Quotient of the Divisor operation
+* Description:  It divides the address by divisor by using bit shift operation
+*               (essentially without explicitely using "/").
+*               Divisor is a power of 2 number and Divided is of u64
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u64 GLOB_u64_Div(u64 addr, u32 divisor)
+{
+       return  (u64)(addr >> GLOB_Calc_Used_Bits(divisor));
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_u64_Remainder
+* Inputs:       Number of u64
+*               Divisor Type (1 -PageAddress, 2- BlockAddress)
+* Outputs:      Remainder of the Division operation
+* Description:  It calculates the remainder of a number (of u64) by
+*               divisor(power of 2 number ) by using bit shifting and multiply
+*               operation(essentially without explicitely using "/").
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u64 GLOB_u64_Remainder(u64 addr, u32 divisor_type)
+{
+       u64 result = 0;
+
+       if (divisor_type == 1) { /* Remainder -- Page */
+               result = (addr >> DeviceInfo.nBitsInPageDataSize);
+               result = result * DeviceInfo.wPageDataSize;
+       } else if (divisor_type == 2) { /* Remainder -- Block */
+               result = (addr >> DeviceInfo.nBitsInBlockDataSize);
+               result = result * DeviceInfo.wBlockDataSize;
+       }
+
+       result = addr - result;
+
+       return result;
+}
+
+#define NUM_DEVICES             1
+#define PARTITIONS              8
+
+#define GLOB_SBD_NAME          "nd"
+#define GLOB_SBD_IRQ_NUM       (29)
+#define GLOB_VERSION           "driver version 20091110"
+
+#define GLOB_SBD_IOCTL_GC                        (0x7701)
+#define GLOB_SBD_IOCTL_WL                        (0x7702)
+#define GLOB_SBD_IOCTL_FORMAT                    (0x7703)
+#define GLOB_SBD_IOCTL_ERASE_FLASH               (0x7704)
+#define GLOB_SBD_IOCTL_FLUSH_CACHE               (0x7705)
+#define GLOB_SBD_IOCTL_COPY_BLK_TABLE            (0x7706)
+#define GLOB_SBD_IOCTL_COPY_WEAR_LEVELING_TABLE  (0x7707)
+#define GLOB_SBD_IOCTL_GET_NAND_INFO             (0x7708)
+#define GLOB_SBD_IOCTL_WRITE_DATA                (0x7709)
+#define GLOB_SBD_IOCTL_READ_DATA                 (0x770A)
+
+static u32 reserved_mb_for_os_image = 0;
+
+int nand_debug_level;
+module_param(nand_debug_level, int, 0644);
+MODULE_PARM_DESC(nand_debug_level, "debug level value: 1-3");
+
+MODULE_LICENSE("GPL");
+
+struct spectra_nand_dev {
+       struct pci_dev *dev;
+       u64 size;
+       u16 users;
+       spinlock_t qlock;
+       void __iomem *ioaddr;  /* Mapped address */
+       struct request_queue *queue;
+       struct task_struct *thread;
+       struct gendisk *gd;
+       u8 *tmp_buf;
+};
+
+
+static int GLOB_SBD_majornum;
+
+static char *GLOB_version = GLOB_VERSION;
+
+static struct spectra_nand_dev nand_device[NUM_DEVICES];
+
+static struct mutex spectra_lock;
+
+static int res_blks_os = 1;
+
+struct spectra_indentfy_dev_tag IdentifyDeviceData;
+
+static int force_flush_cache(void)
+{
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       if (ERR == GLOB_FTL_Flush_Cache()) {
+               printk(KERN_ERR "Fail to Flush FTL Cache!\n");
+               return -EFAULT;
+       }
+#if CMD_DMA
+               if (glob_ftl_execute_cmds())
+                       return -EIO;
+               else
+                       return 0;
+#endif
+       return 0;
+}
+
+struct ioctl_rw_page_info {
+       u8 *data;
+       unsigned int page;
+};
+
+static int ioctl_read_page_data(unsigned long arg)
+{
+       u8 *buf;
+       struct ioctl_rw_page_info info;
+       int result = PASS;
+
+       if (copy_from_user(&info, (void __user *)arg, sizeof(info)))
+               return -EFAULT;
+
+       buf = kmalloc(IdentifyDeviceData.PageDataSize, GFP_ATOMIC);
+       if (!buf) {
+               printk(KERN_ERR "ioctl_read_page_data: "
+                      "failed to allocate memory\n");
+               return -ENOMEM;
+       }
+
+       mutex_lock(&spectra_lock);
+       result = GLOB_FTL_Page_Read(buf,
+               (u64)info.page * IdentifyDeviceData.PageDataSize);
+       mutex_unlock(&spectra_lock);
+
+       if (copy_to_user((void __user *)info.data, buf,
+                          IdentifyDeviceData.PageDataSize)) {
+               printk(KERN_ERR "ioctl_read_page_data: "
+                      "failed to copy user data\n");
+               kfree(buf);
+               return -EFAULT;
+       }
+
+       kfree(buf);
+       return result;
+}
+
+static int ioctl_write_page_data(unsigned long arg)
+{
+       u8 *buf;
+       struct ioctl_rw_page_info info;
+       int result = PASS;
+
+       if (copy_from_user(&info, (void __user *)arg, sizeof(info)))
+               return -EFAULT;
+
+       buf = kmalloc(IdentifyDeviceData.PageDataSize, GFP_ATOMIC);
+       if (!buf) {
+               printk(KERN_ERR "ioctl_write_page_data: "
+                      "failed to allocate memory\n");
+               return -ENOMEM;
+       }
+
+       if (copy_from_user(buf, (void __user *)info.data,
+                          IdentifyDeviceData.PageDataSize)) {
+               printk(KERN_ERR "ioctl_write_page_data: "
+                      "failed to copy user data\n");
+               kfree(buf);
+               return -EFAULT;
+       }
+
+       mutex_lock(&spectra_lock);
+       result = GLOB_FTL_Page_Write(buf,
+               (u64)info.page * IdentifyDeviceData.PageDataSize);
+       mutex_unlock(&spectra_lock);
+
+       kfree(buf);
+       return result;
+}
+
+/* Return how many blocks should be reserved for bad block replacement */
+static int get_res_blk_num_bad_blk(void)
+{
+       return IdentifyDeviceData.wDataBlockNum / 10;
+}
+
+/* Return how many blocks should be reserved for OS image */
+static int get_res_blk_num_os(void)
+{
+       u32 res_blks, blk_size;
+
+       blk_size = IdentifyDeviceData.PageDataSize *
+               IdentifyDeviceData.PagesPerBlock;
+
+       res_blks = (reserved_mb_for_os_image * 1024 * 1024) / blk_size;
+
+       if ((res_blks < 1) || (res_blks >= IdentifyDeviceData.wDataBlockNum))
+               res_blks = 1; /* Reserved 1 block for block table */
+
+       return res_blks;
+}
+
+static void SBD_prepare_flush(struct request_queue *q, struct request *rq)
+{
+       rq->cmd_type = REQ_TYPE_LINUX_BLOCK;
+       /* rq->timeout = 5 * HZ; */
+       rq->cmd[0] = REQ_LB_OP_FLUSH;
+}
+
+/* Transfer a full request. */
+static int do_transfer(struct spectra_nand_dev *tr, struct request *req)
+{
+       u64 start_addr, addr;
+       u32 logical_start_sect, hd_start_sect;
+       u32 nsect, hd_sects;
+       u32 rsect, tsect = 0;
+       char *buf;
+       u32 ratio = IdentifyDeviceData.PageDataSize >> 9;
+
+       start_addr = (u64)(blk_rq_pos(req)) << 9;
+       /* Add a big enough offset to prevent the OS Image from
+       *  being accessed or damaged by file system */
+       start_addr += IdentifyDeviceData.PageDataSize *
+                       IdentifyDeviceData.PagesPerBlock *
+                       res_blks_os;
+
+       if (req->cmd_type == REQ_TYPE_LINUX_BLOCK &&
+                       req->cmd[0] == REQ_LB_OP_FLUSH) {
+               if (force_flush_cache()) /* Fail to flush cache */
+                       return -EIO;
+               else
+                       return 0;
+       }
+
+       if (!blk_fs_request(req))
+               return -EIO;
+
+       if (blk_rq_pos(req) + blk_rq_cur_sectors(req) > get_capacity(tr->gd)) {
+               printk(KERN_ERR "Spectra error: request over the NAND "
+                       "capacity!sector %d, current_nr_sectors %d, "
+                       "while capacity is %d\n",
+                       (int)blk_rq_pos(req),
+                       blk_rq_cur_sectors(req),
+                       (int)get_capacity(tr->gd));
+               return -EIO;
+       }
+
+       logical_start_sect = start_addr >> 9;
+       hd_start_sect = logical_start_sect / ratio;
+       rsect = logical_start_sect - hd_start_sect * ratio;
+
+       addr = (u64)hd_start_sect * ratio * 512;
+       buf = req->buffer;
+       nsect = blk_rq_cur_sectors(req);
+
+       if (rsect)
+               tsect =  (ratio - rsect) < nsect ? (ratio - rsect) : nsect;
+
+       switch (rq_data_dir(req)) {
+       case READ:
+               /* Read the first NAND page */
+               if (rsect) {
+                       if (GLOB_FTL_Page_Read(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       memcpy(buf, tr->tmp_buf + (rsect << 9), tsect << 9);
+                       addr += IdentifyDeviceData.PageDataSize;
+                       buf += tsect << 9;
+                       nsect -= tsect;
+               }
+
+               /* Read the other NAND pages */
+               for (hd_sects = nsect / ratio; hd_sects > 0; hd_sects--) {
+                       if (GLOB_FTL_Page_Read(buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       addr += IdentifyDeviceData.PageDataSize;
+                       buf += IdentifyDeviceData.PageDataSize;
+               }
+
+               /* Read the last NAND pages */
+               if (nsect % ratio) {
+                       if (GLOB_FTL_Page_Read(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       memcpy(buf, tr->tmp_buf, (nsect % ratio) << 9);
+               }
+#if CMD_DMA
+               if (glob_ftl_execute_cmds())
+                       return -EIO;
+               else
+                       return 0;
+#endif
+               return 0;
+
+       case WRITE:
+               /* Write the first NAND page */
+               if (rsect) {
+                       if (GLOB_FTL_Page_Read(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       memcpy(tr->tmp_buf + (rsect << 9), buf, tsect << 9);
+                       if (GLOB_FTL_Page_Write(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       addr += IdentifyDeviceData.PageDataSize;
+                       buf += tsect << 9;
+                       nsect -= tsect;
+               }
+
+               /* Write the other NAND pages */
+               for (hd_sects = nsect / ratio; hd_sects > 0; hd_sects--) {
+                       if (GLOB_FTL_Page_Write(buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       addr += IdentifyDeviceData.PageDataSize;
+                       buf += IdentifyDeviceData.PageDataSize;
+               }
+
+               /* Write the last NAND pages */
+               if (nsect % ratio) {
+                       if (GLOB_FTL_Page_Read(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+                       memcpy(tr->tmp_buf, buf, (nsect % ratio) << 9);
+                       if (GLOB_FTL_Page_Write(tr->tmp_buf, addr)) {
+                               printk(KERN_ERR "Error in %s, Line %d\n",
+                                       __FILE__, __LINE__);
+                               return -EIO;
+                       }
+               }
+#if CMD_DMA
+               if (glob_ftl_execute_cmds())
+                       return -EIO;
+               else
+                       return 0;
+#endif
+               return 0;
+
+       default:
+               printk(KERN_NOTICE "Unknown request %u\n", rq_data_dir(req));
+               return -EIO;
+       }
+}
+
+/* This function is copied from drivers/mtd/mtd_blkdevs.c */
+static int spectra_trans_thread(void *arg)
+{
+       struct spectra_nand_dev *tr = arg;
+       struct request_queue *rq = tr->queue;
+       struct request *req = NULL;
+
+       /* we might get involved when memory gets low, so use PF_MEMALLOC */
+       current->flags |= PF_MEMALLOC;
+
+       spin_lock_irq(rq->queue_lock);
+       while (!kthread_should_stop()) {
+               int res;
+
+               if (!req) {
+                       req = blk_fetch_request(rq);
+                       if (!req) {
+                               set_current_state(TASK_INTERRUPTIBLE);
+                               spin_unlock_irq(rq->queue_lock);
+                               schedule();
+                               spin_lock_irq(rq->queue_lock);
+                               continue;
+                       }
+               }
+
+               spin_unlock_irq(rq->queue_lock);
+
+               mutex_lock(&spectra_lock);
+               res = do_transfer(tr, req);
+               mutex_unlock(&spectra_lock);
+
+               spin_lock_irq(rq->queue_lock);
+
+               if (!__blk_end_request_cur(req, res))
+                       req = NULL;
+       }
+
+       if (req)
+               __blk_end_request_all(req, -EIO);
+
+       spin_unlock_irq(rq->queue_lock);
+
+       return 0;
+}
+
+
+/* Request function that "handles clustering". */
+static void GLOB_SBD_request(struct request_queue *rq)
+{
+       struct spectra_nand_dev *pdev = rq->queuedata;
+       wake_up_process(pdev->thread);
+}
+
+static int GLOB_SBD_open(struct block_device *bdev, fmode_t mode)
+
+{
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+       return 0;
+}
+
+static int GLOB_SBD_release(struct gendisk *disk, fmode_t mode)
+{
+       int ret;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       mutex_lock(&spectra_lock);
+       ret = force_flush_cache();
+       mutex_unlock(&spectra_lock);
+
+       return 0;
+}
+
+static int GLOB_SBD_getgeo(struct block_device *bdev, struct hd_geometry *geo)
+{
+       geo->heads = 4;
+       geo->sectors = 16;
+       geo->cylinders = get_capacity(bdev->bd_disk) / (4 * 16);
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "heads: %d, sectors: %d, cylinders: %d\n",
+               geo->heads, geo->sectors, geo->cylinders);
+
+       return 0;
+}
+
+int GLOB_SBD_ioctl(struct block_device *bdev, fmode_t mode,
+               unsigned int cmd, unsigned long arg)
+{
+       int ret;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       switch (cmd) {
+       case GLOB_SBD_IOCTL_GC:
+               nand_dbg_print(NAND_DBG_DEBUG,
+                              "Spectra IOCTL: Garbage Collection "
+                              "being performed\n");
+               if (PASS != GLOB_FTL_Garbage_Collection())
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_WL:
+               nand_dbg_print(NAND_DBG_DEBUG,
+                              "Spectra IOCTL: Static Wear Leveling "
+                              "being performed\n");
+               if (PASS != GLOB_FTL_Wear_Leveling())
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_FORMAT:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: Flash format "
+                              "being performed\n");
+               if (PASS != GLOB_FTL_Flash_Format())
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_FLUSH_CACHE:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: Cache flush "
+                              "being performed\n");
+               mutex_lock(&spectra_lock);
+               ret = force_flush_cache();
+               mutex_unlock(&spectra_lock);
+               return ret;
+
+       case GLOB_SBD_IOCTL_COPY_BLK_TABLE:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: "
+                              "Copy block table\n");
+               if (copy_to_user((void __user *)arg,
+                       get_blk_table_start_addr(),
+                       get_blk_table_len()))
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_COPY_WEAR_LEVELING_TABLE:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: "
+                              "Copy wear leveling table\n");
+               if (copy_to_user((void __user *)arg,
+                       get_wear_leveling_table_start_addr(),
+                       get_wear_leveling_table_len()))
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_GET_NAND_INFO:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: "
+                              "Get NAND info\n");
+               if (copy_to_user((void __user *)arg, &IdentifyDeviceData,
+                       sizeof(IdentifyDeviceData)))
+                       return -EFAULT;
+               return 0;
+
+       case GLOB_SBD_IOCTL_WRITE_DATA:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: "
+                              "Write one page data\n");
+               return ioctl_write_page_data(arg);
+
+       case GLOB_SBD_IOCTL_READ_DATA:
+               nand_dbg_print(NAND_DBG_DEBUG, "Spectra IOCTL: "
+                              "Read one page data\n");
+               return ioctl_read_page_data(arg);
+       }
+
+       return -ENOTTY;
+}
+
+static struct block_device_operations GLOB_SBD_ops = {
+       .owner = THIS_MODULE,
+       .open = GLOB_SBD_open,
+       .release = GLOB_SBD_release,
+       .locked_ioctl = GLOB_SBD_ioctl,
+       .getgeo = GLOB_SBD_getgeo,
+};
+
+static int SBD_setup_device(struct spectra_nand_dev *dev, int which)
+{
+       int res_blks;
+       u32 sects;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       memset(dev, 0, sizeof(struct spectra_nand_dev));
+
+       nand_dbg_print(NAND_DBG_WARN, "Reserved %d blocks "
+               "for OS image, %d blocks for bad block replacement.\n",
+               get_res_blk_num_os(),
+               get_res_blk_num_bad_blk());
+
+       res_blks = get_res_blk_num_bad_blk() + get_res_blk_num_os();
+
+       dev->size = (u64)IdentifyDeviceData.PageDataSize *
+               IdentifyDeviceData.PagesPerBlock *
+               (IdentifyDeviceData.wDataBlockNum - res_blks);
+
+       res_blks_os = get_res_blk_num_os();
+
+       spin_lock_init(&dev->qlock);
+
+       dev->tmp_buf = kmalloc(IdentifyDeviceData.PageDataSize, GFP_ATOMIC);
+       if (!dev->tmp_buf) {
+               printk(KERN_ERR "Failed to kmalloc memory in %s Line %d, exit.\n",
+                       __FILE__, __LINE__);
+               goto out_vfree;
+       }
+
+       dev->queue = blk_init_queue(GLOB_SBD_request, &dev->qlock);
+       if (dev->queue == NULL) {
+               printk(KERN_ERR
+                      "Spectra: Request queue could not be initialized."
+                       " Aborting\n ");
+               goto out_vfree;
+       }
+       dev->queue->queuedata = dev;
+
+       /* As Linux block layer doens't support >4KB hardware sector,  */
+       /* Here we force report 512 byte hardware sector size to Kernel */
+       blk_queue_logical_block_size(dev->queue, 512);
+
+       blk_queue_ordered(dev->queue, QUEUE_ORDERED_DRAIN_FLUSH,
+                                       SBD_prepare_flush);
+
+       dev->thread = kthread_run(spectra_trans_thread, dev, "nand_thd");
+       if (IS_ERR(dev->thread)) {
+               blk_cleanup_queue(dev->queue);
+               unregister_blkdev(GLOB_SBD_majornum, GLOB_SBD_NAME);
+               return PTR_ERR(dev->thread);
+       }
+
+       dev->gd = alloc_disk(PARTITIONS);
+       if (!dev->gd) {
+               printk(KERN_ERR
+                      "Spectra: Could not allocate disk. Aborting \n ");
+               goto out_vfree;
+       }
+       dev->gd->major = GLOB_SBD_majornum;
+       dev->gd->first_minor = which * PARTITIONS;
+       dev->gd->fops = &GLOB_SBD_ops;
+       dev->gd->queue = dev->queue;
+       dev->gd->private_data = dev;
+       snprintf(dev->gd->disk_name, 32, "%s%c", GLOB_SBD_NAME, which + 'a');
+
+       sects = dev->size >> 9;
+       nand_dbg_print(NAND_DBG_WARN, "Capacity sects: %d\n", sects);
+       set_capacity(dev->gd, sects);
+
+       add_disk(dev->gd);
+
+       return 0;
+out_vfree:
+       return -ENOMEM;
+}
+
+/*
+static ssize_t show_nand_block_num(struct device *dev,
+       struct device_attribute *attr, char *buf)
+{
+       return snprintf(buf, PAGE_SIZE, "%d\n",
+               (int)IdentifyDeviceData.wDataBlockNum);
+}
+
+static ssize_t show_nand_pages_per_block(struct device *dev,
+       struct device_attribute *attr, char *buf)
+{
+       return snprintf(buf, PAGE_SIZE, "%d\n",
+               (int)IdentifyDeviceData.PagesPerBlock);
+}
+
+static ssize_t show_nand_page_size(struct device *dev,
+       struct device_attribute *attr, char *buf)
+{
+       return snprintf(buf, PAGE_SIZE, "%d\n",
+               (int)IdentifyDeviceData.PageDataSize);
+}
+
+static DEVICE_ATTR(nand_block_num, 0444, show_nand_block_num, NULL);
+static DEVICE_ATTR(nand_pages_per_block, 0444, show_nand_pages_per_block, NULL);
+static DEVICE_ATTR(nand_page_size, 0444, show_nand_page_size, NULL);
+
+static void create_sysfs_entry(struct device *dev)
+{
+       if (device_create_file(dev, &dev_attr_nand_block_num))
+               printk(KERN_ERR "Spectra: "
+                       "failed to create sysfs entry nand_block_num.\n");
+       if (device_create_file(dev, &dev_attr_nand_pages_per_block))
+               printk(KERN_ERR "Spectra: "
+               "failed to create sysfs entry nand_pages_per_block.\n");
+       if (device_create_file(dev, &dev_attr_nand_page_size))
+               printk(KERN_ERR "Spectra: "
+               "failed to create sysfs entry nand_page_size.\n");
+}
+*/
+
+static int GLOB_SBD_init(void)
+{
+       int i;
+
+       /* Set debug output level (0~3) here. 3 is most verbose */
+       nand_debug_level = 0;
+
+       printk(KERN_ALERT "Spectra: %s\n", GLOB_version);
+
+       mutex_init(&spectra_lock);
+
+       GLOB_SBD_majornum = register_blkdev(0, GLOB_SBD_NAME);
+       if (GLOB_SBD_majornum <= 0) {
+               printk(KERN_ERR "Unable to get the major %d for Spectra",
+                      GLOB_SBD_majornum);
+               return -EBUSY;
+       }
+
+       if (PASS != GLOB_FTL_Flash_Init()) {
+               printk(KERN_ERR "Spectra: Unable to Initialize Flash Device. "
+                      "Aborting\n");
+               goto out_flash_register;
+       }
+
+       /* create_sysfs_entry(&dev->dev); */
+
+       if (PASS != GLOB_FTL_IdentifyDevice(&IdentifyDeviceData)) {
+               printk(KERN_ERR "Spectra: Unable to Read Flash Device. "
+                      "Aborting\n");
+               goto out_flash_register;
+       } else {
+               nand_dbg_print(NAND_DBG_WARN, "In GLOB_SBD_init: "
+                              "Num blocks=%d, pagesperblock=%d, "
+                              "pagedatasize=%d, ECCBytesPerSector=%d\n",
+                      (int)IdentifyDeviceData.NumBlocks,
+                      (int)IdentifyDeviceData.PagesPerBlock,
+                      (int)IdentifyDeviceData.PageDataSize,
+                      (int)IdentifyDeviceData.wECCBytesPerSector);
+       }
+
+       printk(KERN_ALERT "Spectra: searching block table, please wait ...\n");
+       if (GLOB_FTL_Init() != PASS) {
+               printk(KERN_ERR "Spectra: Unable to Initialize FTL Layer. "
+                      "Aborting\n");
+               goto out_ftl_flash_register;
+       }
+       printk(KERN_ALERT "Spectra: block table has been found.\n");
+
+       for (i = 0; i < NUM_DEVICES; i++)
+               if (SBD_setup_device(&nand_device[i], i) == -ENOMEM)
+                       goto out_ftl_flash_register;
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+                      "Spectra: module loaded with major number %d\n",
+                      GLOB_SBD_majornum);
+
+       return 0;
+
+out_ftl_flash_register:
+       GLOB_FTL_Cache_Release();
+out_flash_register:
+       GLOB_FTL_Flash_Release();
+       unregister_blkdev(GLOB_SBD_majornum, GLOB_SBD_NAME);
+       printk(KERN_ERR "Spectra: Module load failed.\n");
+
+       return -ENOMEM;
+}
+
+static void __exit GLOB_SBD_exit(void)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < NUM_DEVICES; i++) {
+               struct spectra_nand_dev *dev = &nand_device[i];
+               if (dev->gd) {
+                       del_gendisk(dev->gd);
+                       put_disk(dev->gd);
+               }
+               if (dev->queue)
+                       blk_cleanup_queue(dev->queue);
+               kfree(dev->tmp_buf);
+       }
+
+       unregister_blkdev(GLOB_SBD_majornum, GLOB_SBD_NAME);
+
+       mutex_lock(&spectra_lock);
+       force_flush_cache();
+       mutex_unlock(&spectra_lock);
+
+       GLOB_FTL_Cache_Release();
+
+       GLOB_FTL_Flash_Release();
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+                      "Spectra FTL module (major number %d) unloaded.\n",
+                      GLOB_SBD_majornum);
+}
+
+static int __init setup_reserve_space_for_os_image(char *cmdline)
+{
+       unsigned long value;
+       int error;
+
+       printk(KERN_ALERT "Spectra - cmdline: %s\n", cmdline);
+       if (!cmdline)
+               return -EINVAL;
+
+       error = strict_strtoul((const char *)cmdline, 10, &value);
+       if (error)
+               return -EINVAL;
+
+       reserved_mb_for_os_image = value;
+
+       return 0;
+}
+
+early_param("res_nand", setup_reserve_space_for_os_image);
+
+module_init(GLOB_SBD_init);
+module_exit(GLOB_SBD_exit);
diff --git a/drivers/block/spectra/ffsport.h b/drivers/block/spectra/ffsport.h
new file mode 100644 (file)
index 0000000..6c5d90c
--- /dev/null
@@ -0,0 +1,84 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _FFSPORT_
+#define _FFSPORT_
+
+#include "ffsdefs.h"
+
+#if defined __GNUC__
+#define PACKED
+#define PACKED_GNU __attribute__ ((packed))
+#define UNALIGNED
+#endif
+
+#include <linux/semaphore.h>
+#include <linux/string.h>      /* for strcpy(), stricmp(), etc */
+#include <linux/mm.h>          /* for kmalloc(), kfree() */
+#include <linux/vmalloc.h>
+#include <linux/module.h>
+#include <linux/moduleparam.h>
+#include <linux/init.h>
+
+#include <linux/kernel.h>      /* printk() */
+#include <linux/fs.h>          /* everything... */
+#include <linux/errno.h>       /* error codes */
+#include <linux/types.h>       /* size_t */
+#include <linux/genhd.h>
+#include <linux/blkdev.h>
+#include <linux/hdreg.h>
+#include <linux/pci.h>
+#include "flash.h"
+
+#define VERBOSE    1
+
+#define NAND_DBG_WARN  1
+#define NAND_DBG_DEBUG 2
+#define NAND_DBG_TRACE 3
+
+extern int nand_debug_level;
+
+#ifdef VERBOSE
+#define nand_dbg_print(level, args...)                 \
+       do {                                            \
+               if (level <= nand_debug_level)          \
+                       printk(KERN_ALERT args);        \
+       } while (0)
+#else
+#define nand_dbg_print(level, args...)
+#endif
+
+#ifdef SUPPORT_BIG_ENDIAN
+#define INVERTUINT16(w)   ((u16)(((u16)(w)) << 8) | \
+                          (u16)((u16)(w) >> 8))
+
+#define INVERTUINT32(dw)  (((u32)(dw) << 24) | \
+                          (((u32)(dw) << 8) & 0x00ff0000) | \
+                          (((u32)(dw) >> 8) & 0x0000ff00) | \
+                          ((u32)(dw) >> 24))
+#else
+#define INVERTUINT16(w)   w
+#define INVERTUINT32(dw)  dw
+#endif
+
+extern int GLOB_Calc_Used_Bits(u32 n);
+extern u64 GLOB_u64_Div(u64 addr, u32 divisor);
+extern u64 GLOB_u64_Remainder(u64 addr, u32 divisor_type);
+
+#endif /* _FFSPORT_ */
diff --git a/drivers/block/spectra/flash.c b/drivers/block/spectra/flash.c
new file mode 100644 (file)
index 0000000..134aa51
--- /dev/null
@@ -0,0 +1,4731 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include <linux/fs.h>
+#include <linux/slab.h>
+
+#include "flash.h"
+#include "ffsdefs.h"
+#include "lld.h"
+#include "lld_nand.h"
+#if CMD_DMA
+#include "lld_cdma.h"
+#endif
+
+#define BLK_FROM_ADDR(addr)  ((u32)(addr >> DeviceInfo.nBitsInBlockDataSize))
+#define PAGE_FROM_ADDR(addr, Block)  ((u16)((addr - (u64)Block * \
+       DeviceInfo.wBlockDataSize) >> DeviceInfo.nBitsInPageDataSize))
+
+#define IS_SPARE_BLOCK(blk)     (BAD_BLOCK != (pbt[blk] &\
+       BAD_BLOCK) && SPARE_BLOCK == (pbt[blk] & SPARE_BLOCK))
+
+#define IS_DATA_BLOCK(blk)      (0 == (pbt[blk] & BAD_BLOCK))
+
+#define IS_DISCARDED_BLOCK(blk) (BAD_BLOCK != (pbt[blk] &\
+       BAD_BLOCK) && DISCARD_BLOCK == (pbt[blk] & DISCARD_BLOCK))
+
+#define IS_BAD_BLOCK(blk)       (BAD_BLOCK == (pbt[blk] & BAD_BLOCK))
+
+#if DEBUG_BNDRY
+void debug_boundary_lineno_error(int chnl, int limit, int no,
+                               int lineno, char *filename)
+{
+       if (chnl >= limit)
+               printk(KERN_ERR "Boundary Check Fail value %d >= limit %d, "
+               "at  %s:%d. Other info:%d. Aborting...\n",
+               chnl, limit, filename, lineno, no);
+}
+/* static int globalmemsize; */
+#endif
+
+static u16 FTL_Cache_If_Hit(u64 dwPageAddr);
+static int FTL_Cache_Read(u64 dwPageAddr);
+static void FTL_Cache_Read_Page(u8 *pData, u64 dwPageAddr,
+                               u16 cache_blk);
+static void FTL_Cache_Write_Page(u8 *pData, u64 dwPageAddr,
+                                u8 cache_blk, u16 flag);
+static int FTL_Cache_Write(void);
+static int FTL_Cache_Write_Back(u8 *pData, u64 blk_addr);
+static void FTL_Calculate_LRU(void);
+static u32 FTL_Get_Block_Index(u32 wBlockNum);
+
+static int FTL_Search_Block_Table_IN_Block(u32 BT_Block,
+                                          u8 BT_Tag, u16 *Page);
+static int FTL_Read_Block_Table(void);
+static int FTL_Write_Block_Table(int wForce);
+static int FTL_Write_Block_Table_Data(void);
+static int FTL_Check_Block_Table(int wOldTable);
+static int FTL_Static_Wear_Leveling(void);
+static u32 FTL_Replace_Block_Table(void);
+static int FTL_Write_IN_Progress_Block_Table_Page(void);
+
+static u32 FTL_Get_Page_Num(u64 length);
+static u64 FTL_Get_Physical_Block_Addr(u64 blk_addr);
+
+static u32 FTL_Replace_OneBlock(u32 wBlockNum,
+                                     u32 wReplaceNum);
+static u32 FTL_Replace_LWBlock(u32 wBlockNum,
+                                    int *pGarbageCollect);
+static u32 FTL_Replace_MWBlock(void);
+static int FTL_Replace_Block(u64 blk_addr);
+static int FTL_Adjust_Relative_Erase_Count(u32 Index_of_MAX);
+
+static int FTL_Flash_Error_Handle(u8 *pData, u64 old_page_addr, u64 blk_addr);
+
+struct device_info_tag DeviceInfo;
+struct flash_cache_tag Cache;
+static struct spectra_l2_cache_info cache_l2;
+
+static u8 *cache_l2_page_buf;
+static u8 *cache_l2_blk_buf;
+
+u8 *g_pBlockTable;
+u8 *g_pWearCounter;
+u16 *g_pReadCounter;
+u32 *g_pBTBlocks;
+static u16 g_wBlockTableOffset;
+static u32 g_wBlockTableIndex;
+static u8 g_cBlockTableStatus;
+
+static u8 *g_pTempBuf;
+static u8 *flag_check_blk_table;
+static u8 *tmp_buf_search_bt_in_block;
+static u8 *spare_buf_search_bt_in_block;
+static u8 *spare_buf_bt_search_bt_in_block;
+static u8 *tmp_buf1_read_blk_table;
+static u8 *tmp_buf2_read_blk_table;
+static u8 *flags_static_wear_leveling;
+static u8 *tmp_buf_write_blk_table_data;
+static u8 *tmp_buf_read_disturbance;
+
+u8 *buf_read_page_main_spare;
+u8 *buf_write_page_main_spare;
+u8 *buf_read_page_spare;
+u8 *buf_get_bad_block;
+
+#if (RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE && CMD_DMA)
+struct flash_cache_delta_list_tag int_cache[MAX_CHANS + MAX_DESCS];
+struct flash_cache_tag cache_start_copy;
+#endif
+
+int g_wNumFreeBlocks;
+u8 g_SBDCmdIndex;
+
+static u8 *g_pIPF;
+static u8 bt_flag = FIRST_BT_ID;
+static u8 bt_block_changed;
+
+static u16 cache_block_to_write;
+static u8 last_erased = FIRST_BT_ID;
+
+static u8 GC_Called;
+static u8 BT_GC_Called;
+
+#if CMD_DMA
+#define COPY_BACK_BUF_NUM 10
+
+static u8 ftl_cmd_cnt;  /* Init value is 0 */
+u8 *g_pBTDelta;
+u8 *g_pBTDelta_Free;
+u8 *g_pBTStartingCopy;
+u8 *g_pWearCounterCopy;
+u16 *g_pReadCounterCopy;
+u8 *g_pBlockTableCopies;
+u8 *g_pNextBlockTable;
+static u8 *cp_back_buf_copies[COPY_BACK_BUF_NUM];
+static int cp_back_buf_idx;
+
+static u8 *g_temp_buf;
+
+#pragma pack(push, 1)
+#pragma pack(1)
+struct BTableChangesDelta {
+       u8 ftl_cmd_cnt;
+       u8 ValidFields;
+       u16 g_wBlockTableOffset;
+       u32 g_wBlockTableIndex;
+       u32 BT_Index;
+       u32 BT_Entry_Value;
+       u32 WC_Index;
+       u8 WC_Entry_Value;
+       u32 RC_Index;
+       u16 RC_Entry_Value;
+};
+
+#pragma pack(pop)
+
+struct BTableChangesDelta *p_BTableChangesDelta;
+#endif
+
+
+#define MARK_BLOCK_AS_BAD(blocknode)      (blocknode |= BAD_BLOCK)
+#define MARK_BLK_AS_DISCARD(blk)  (blk = (blk & ~SPARE_BLOCK) | DISCARD_BLOCK)
+
+#define FTL_Get_LBAPBA_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
+                                               sizeof(u32))
+#define FTL_Get_WearCounter_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
+                                               sizeof(u8))
+#define FTL_Get_ReadCounter_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
+                                               sizeof(u16))
+#if SUPPORT_LARGE_BLOCKNUM
+#define FTL_Get_LBAPBA_Table_Flash_Size_Bytes() (DeviceInfo.wDataBlockNum *\
+                                               sizeof(u8) * 3)
+#else
+#define FTL_Get_LBAPBA_Table_Flash_Size_Bytes() (DeviceInfo.wDataBlockNum *\
+                                               sizeof(u16))
+#endif
+#define FTL_Get_WearCounter_Table_Flash_Size_Bytes \
+       FTL_Get_WearCounter_Table_Mem_Size_Bytes
+#define FTL_Get_ReadCounter_Table_Flash_Size_Bytes \
+       FTL_Get_ReadCounter_Table_Mem_Size_Bytes
+
+static u32 FTL_Get_Block_Table_Flash_Size_Bytes(void)
+{
+       u32 byte_num;
+
+       if (DeviceInfo.MLCDevice) {
+               byte_num = FTL_Get_LBAPBA_Table_Flash_Size_Bytes() +
+                       DeviceInfo.wDataBlockNum * sizeof(u8) +
+                       DeviceInfo.wDataBlockNum * sizeof(u16);
+       } else {
+               byte_num = FTL_Get_LBAPBA_Table_Flash_Size_Bytes() +
+                       DeviceInfo.wDataBlockNum * sizeof(u8);
+       }
+
+       byte_num += 4 * sizeof(u8);
+
+       return byte_num;
+}
+
+static u16  FTL_Get_Block_Table_Flash_Size_Pages(void)
+{
+       return (u16)FTL_Get_Page_Num(FTL_Get_Block_Table_Flash_Size_Bytes());
+}
+
+static int FTL_Copy_Block_Table_To_Flash(u8 *flashBuf, u32 sizeToTx,
+                                       u32 sizeTxed)
+{
+       u32 wBytesCopied, blk_tbl_size, wBytes;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       blk_tbl_size = FTL_Get_LBAPBA_Table_Flash_Size_Bytes();
+       for (wBytes = 0;
+       (wBytes < sizeToTx) && ((wBytes + sizeTxed) < blk_tbl_size);
+       wBytes++) {
+#if SUPPORT_LARGE_BLOCKNUM
+               flashBuf[wBytes] = (u8)(pbt[(wBytes + sizeTxed) / 3]
+               >> (((wBytes + sizeTxed) % 3) ?
+               ((((wBytes + sizeTxed) % 3) == 2) ? 0 : 8) : 16)) & 0xFF;
+#else
+               flashBuf[wBytes] = (u8)(pbt[(wBytes + sizeTxed) / 2]
+               >> (((wBytes + sizeTxed) % 2) ? 0 : 8)) & 0xFF;
+#endif
+       }
+
+       sizeTxed = (sizeTxed > blk_tbl_size) ? (sizeTxed - blk_tbl_size) : 0;
+       blk_tbl_size = FTL_Get_WearCounter_Table_Flash_Size_Bytes();
+       wBytesCopied = wBytes;
+       wBytes = ((blk_tbl_size - sizeTxed) > (sizeToTx - wBytesCopied)) ?
+               (sizeToTx - wBytesCopied) : (blk_tbl_size - sizeTxed);
+       memcpy(flashBuf + wBytesCopied, g_pWearCounter + sizeTxed, wBytes);
+
+       sizeTxed = (sizeTxed > blk_tbl_size) ? (sizeTxed - blk_tbl_size) : 0;
+
+       if (DeviceInfo.MLCDevice) {
+               blk_tbl_size = FTL_Get_ReadCounter_Table_Flash_Size_Bytes();
+               wBytesCopied += wBytes;
+               for (wBytes = 0; ((wBytes + wBytesCopied) < sizeToTx) &&
+                       ((wBytes + sizeTxed) < blk_tbl_size); wBytes++)
+                       flashBuf[wBytes + wBytesCopied] =
+                       (g_pReadCounter[(wBytes + sizeTxed) / 2] >>
+                       (((wBytes + sizeTxed) % 2) ? 0 : 8)) & 0xFF;
+       }
+
+       return wBytesCopied + wBytes;
+}
+
+static int FTL_Copy_Block_Table_From_Flash(u8 *flashBuf,
+                               u32 sizeToTx, u32 sizeTxed)
+{
+       u32 wBytesCopied, blk_tbl_size, wBytes;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       blk_tbl_size = FTL_Get_LBAPBA_Table_Flash_Size_Bytes();
+       for (wBytes = 0; (wBytes < sizeToTx) &&
+               ((wBytes + sizeTxed) < blk_tbl_size); wBytes++) {
+#if SUPPORT_LARGE_BLOCKNUM
+               if (!((wBytes + sizeTxed) % 3))
+                       pbt[(wBytes + sizeTxed) / 3] = 0;
+               pbt[(wBytes + sizeTxed) / 3] |=
+                       (flashBuf[wBytes] << (((wBytes + sizeTxed) % 3) ?
+                       ((((wBytes + sizeTxed) % 3) == 2) ? 0 : 8) : 16));
+#else
+               if (!((wBytes + sizeTxed) % 2))
+                       pbt[(wBytes + sizeTxed) / 2] = 0;
+               pbt[(wBytes + sizeTxed) / 2] |=
+                       (flashBuf[wBytes] << (((wBytes + sizeTxed) % 2) ?
+                       0 : 8));
+#endif
+       }
+
+       sizeTxed = (sizeTxed > blk_tbl_size) ? (sizeTxed - blk_tbl_size) : 0;
+       blk_tbl_size = FTL_Get_WearCounter_Table_Flash_Size_Bytes();
+       wBytesCopied = wBytes;
+       wBytes = ((blk_tbl_size - sizeTxed) > (sizeToTx - wBytesCopied)) ?
+               (sizeToTx - wBytesCopied) : (blk_tbl_size - sizeTxed);
+       memcpy(g_pWearCounter + sizeTxed, flashBuf + wBytesCopied, wBytes);
+       sizeTxed = (sizeTxed > blk_tbl_size) ? (sizeTxed - blk_tbl_size) : 0;
+
+       if (DeviceInfo.MLCDevice) {
+               wBytesCopied += wBytes;
+               blk_tbl_size = FTL_Get_ReadCounter_Table_Flash_Size_Bytes();
+               for (wBytes = 0; ((wBytes + wBytesCopied) < sizeToTx) &&
+                       ((wBytes + sizeTxed) < blk_tbl_size); wBytes++) {
+                       if (((wBytes + sizeTxed) % 2))
+                               g_pReadCounter[(wBytes + sizeTxed) / 2] = 0;
+                       g_pReadCounter[(wBytes + sizeTxed) / 2] |=
+                               (flashBuf[wBytes] <<
+                               (((wBytes + sizeTxed) % 2) ? 0 : 8));
+               }
+       }
+
+       return wBytesCopied+wBytes;
+}
+
+static int FTL_Insert_Block_Table_Signature(u8 *buf, u8 tag)
+{
+       int i;
+
+       for (i = 0; i < BTSIG_BYTES; i++)
+               buf[BTSIG_OFFSET + i] =
+               ((tag + (i * BTSIG_DELTA) - FIRST_BT_ID) %
+               (1 + LAST_BT_ID-FIRST_BT_ID)) + FIRST_BT_ID;
+
+       return PASS;
+}
+
+static int FTL_Extract_Block_Table_Tag(u8 *buf, u8 **tagarray)
+{
+       static u8 tag[BTSIG_BYTES >> 1];
+       int i, j, k, tagi, tagtemp, status;
+
+       *tagarray = (u8 *)tag;
+       tagi = 0;
+
+       for (i = 0; i < (BTSIG_BYTES - 1); i++) {
+               for (j = i + 1; (j < BTSIG_BYTES) &&
+                       (tagi < (BTSIG_BYTES >> 1)); j++) {
+                       tagtemp = buf[BTSIG_OFFSET + j] -
+                               buf[BTSIG_OFFSET + i];
+                       if (tagtemp && !(tagtemp % BTSIG_DELTA)) {
+                               tagtemp = (buf[BTSIG_OFFSET + i] +
+                                       (1 + LAST_BT_ID - FIRST_BT_ID) -
+                                       (i * BTSIG_DELTA)) %
+                                       (1 + LAST_BT_ID - FIRST_BT_ID);
+                               status = FAIL;
+                               for (k = 0; k < tagi; k++) {
+                                       if (tagtemp == tag[k])
+                                               status = PASS;
+                               }
+
+                               if (status == FAIL) {
+                                       tag[tagi++] = tagtemp;
+                                       i = (j == (i + 1)) ? i + 1 : i;
+                                       j = (j == (i + 1)) ? i + 1 : i;
+                               }
+                       }
+               }
+       }
+
+       return tagi;
+}
+
+
+static int FTL_Execute_SPL_Recovery(void)
+{
+       u32 j, block, blks;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       int ret;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                               __FILE__, __LINE__, __func__);
+
+       blks = DeviceInfo.wSpectraEndBlock - DeviceInfo.wSpectraStartBlock;
+       for (j = 0; j <= blks; j++) {
+               block = (pbt[j]);
+               if (((block & BAD_BLOCK) != BAD_BLOCK) &&
+                       ((block & SPARE_BLOCK) == SPARE_BLOCK)) {
+                       ret =  GLOB_LLD_Erase_Block(block & ~BAD_BLOCK);
+                       if (FAIL == ret) {
+                               nand_dbg_print(NAND_DBG_WARN,
+                                       "NAND Program fail in %s, Line %d, "
+                                       "Function: %s, new Bad Block %d "
+                                       "generated!\n",
+                                       __FILE__, __LINE__, __func__,
+                                       (int)(block & ~BAD_BLOCK));
+                               MARK_BLOCK_AS_BAD(pbt[j]);
+                       }
+               }
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_IdentifyDevice
+* Inputs:       pointer to identify data structure
+* Outputs:      PASS / FAIL
+* Description:  the identify data structure is filled in with
+*                   information for the block driver.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_IdentifyDevice(struct spectra_indentfy_dev_tag *dev_data)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                               __FILE__, __LINE__, __func__);
+
+       dev_data->NumBlocks = DeviceInfo.wTotalBlocks;
+       dev_data->PagesPerBlock = DeviceInfo.wPagesPerBlock;
+       dev_data->PageDataSize = DeviceInfo.wPageDataSize;
+       dev_data->wECCBytesPerSector = DeviceInfo.wECCBytesPerSector;
+       dev_data->wDataBlockNum = DeviceInfo.wDataBlockNum;
+
+       return PASS;
+}
+
+/* ..... */
+static int allocate_memory(void)
+{
+       u32 block_table_size, page_size, block_size, mem_size;
+       u32 total_bytes = 0;
+       int i;
+#if CMD_DMA
+       int j;
+#endif
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       page_size = DeviceInfo.wPageSize;
+       block_size = DeviceInfo.wPagesPerBlock * DeviceInfo.wPageDataSize;
+
+       block_table_size = DeviceInfo.wDataBlockNum *
+               (sizeof(u32) + sizeof(u8) + sizeof(u16));
+       block_table_size += (DeviceInfo.wPageDataSize -
+               (block_table_size % DeviceInfo.wPageDataSize)) %
+               DeviceInfo.wPageDataSize;
+
+       /* Malloc memory for block tables */
+       g_pBlockTable = kmalloc(block_table_size, GFP_ATOMIC);
+       if (!g_pBlockTable)
+               goto block_table_fail;
+       memset(g_pBlockTable, 0, block_table_size);
+       total_bytes += block_table_size;
+
+       g_pWearCounter = (u8 *)(g_pBlockTable +
+               DeviceInfo.wDataBlockNum * sizeof(u32));
+
+       if (DeviceInfo.MLCDevice)
+               g_pReadCounter = (u16 *)(g_pBlockTable +
+                       DeviceInfo.wDataBlockNum *
+                       (sizeof(u32) + sizeof(u8)));
+
+       /* Malloc memory and init for cache items */
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               Cache.array[i].address = NAND_CACHE_INIT_ADDR;
+               Cache.array[i].use_cnt = 0;
+               Cache.array[i].changed = CLEAR;
+               Cache.array[i].buf = kmalloc(Cache.cache_item_size,
+                       GFP_ATOMIC);
+               if (!Cache.array[i].buf)
+                       goto cache_item_fail;
+               memset(Cache.array[i].buf, 0, Cache.cache_item_size);
+               total_bytes += Cache.cache_item_size;
+       }
+
+       /* Malloc memory for IPF */
+       g_pIPF = kmalloc(page_size, GFP_ATOMIC);
+       if (!g_pIPF)
+               goto ipf_fail;
+       memset(g_pIPF, 0, page_size);
+       total_bytes += page_size;
+
+       /* Malloc memory for data merging during Level2 Cache flush */
+       cache_l2_page_buf = kmalloc(page_size, GFP_ATOMIC);
+       if (!cache_l2_page_buf)
+               goto cache_l2_page_buf_fail;
+       memset(cache_l2_page_buf, 0xff, page_size);
+       total_bytes += page_size;
+
+       cache_l2_blk_buf = kmalloc(block_size, GFP_ATOMIC);
+       if (!cache_l2_blk_buf)
+               goto cache_l2_blk_buf_fail;
+       memset(cache_l2_blk_buf, 0xff, block_size);
+       total_bytes += block_size;
+
+       /* Malloc memory for temp buffer */
+       g_pTempBuf = kmalloc(Cache.cache_item_size, GFP_ATOMIC);
+       if (!g_pTempBuf)
+               goto Temp_buf_fail;
+       memset(g_pTempBuf, 0, Cache.cache_item_size);
+       total_bytes += Cache.cache_item_size;
+
+       /* Malloc memory for block table blocks */
+       mem_size = (1 + LAST_BT_ID - FIRST_BT_ID) * sizeof(u32);
+       g_pBTBlocks = kmalloc(mem_size, GFP_ATOMIC);
+       if (!g_pBTBlocks)
+               goto bt_blocks_fail;
+       memset(g_pBTBlocks, 0xff, mem_size);
+       total_bytes += mem_size;
+
+       /* Malloc memory for function FTL_Check_Block_Table */
+       flag_check_blk_table = kmalloc(DeviceInfo.wDataBlockNum, GFP_ATOMIC);
+       if (!flag_check_blk_table)
+               goto flag_check_blk_table_fail;
+       total_bytes += DeviceInfo.wDataBlockNum;
+
+       /* Malloc memory for function FTL_Search_Block_Table_IN_Block */
+       tmp_buf_search_bt_in_block = kmalloc(page_size, GFP_ATOMIC);
+       if (!tmp_buf_search_bt_in_block)
+               goto tmp_buf_search_bt_in_block_fail;
+       memset(tmp_buf_search_bt_in_block, 0xff, page_size);
+       total_bytes += page_size;
+
+       mem_size = DeviceInfo.wPageSize - DeviceInfo.wPageDataSize;
+       spare_buf_search_bt_in_block = kmalloc(mem_size, GFP_ATOMIC);
+       if (!spare_buf_search_bt_in_block)
+               goto spare_buf_search_bt_in_block_fail;
+       memset(spare_buf_search_bt_in_block, 0xff, mem_size);
+       total_bytes += mem_size;
+
+       spare_buf_bt_search_bt_in_block = kmalloc(mem_size, GFP_ATOMIC);
+       if (!spare_buf_bt_search_bt_in_block)
+               goto spare_buf_bt_search_bt_in_block_fail;
+       memset(spare_buf_bt_search_bt_in_block, 0xff, mem_size);
+       total_bytes += mem_size;
+
+       /* Malloc memory for function FTL_Read_Block_Table */
+       tmp_buf1_read_blk_table = kmalloc(page_size, GFP_ATOMIC);
+       if (!tmp_buf1_read_blk_table)
+               goto tmp_buf1_read_blk_table_fail;
+       memset(tmp_buf1_read_blk_table, 0xff, page_size);
+       total_bytes += page_size;
+
+       tmp_buf2_read_blk_table = kmalloc(page_size, GFP_ATOMIC);
+       if (!tmp_buf2_read_blk_table)
+               goto tmp_buf2_read_blk_table_fail;
+       memset(tmp_buf2_read_blk_table, 0xff, page_size);
+       total_bytes += page_size;
+
+       /* Malloc memory for function FTL_Static_Wear_Leveling */
+       flags_static_wear_leveling = kmalloc(DeviceInfo.wDataBlockNum,
+                                       GFP_ATOMIC);
+       if (!flags_static_wear_leveling)
+               goto flags_static_wear_leveling_fail;
+       total_bytes += DeviceInfo.wDataBlockNum;
+
+       /* Malloc memory for function FTL_Write_Block_Table_Data */
+       if (FTL_Get_Block_Table_Flash_Size_Pages() > 3)
+               mem_size = FTL_Get_Block_Table_Flash_Size_Bytes() -
+                               2 * DeviceInfo.wPageSize;
+       else
+               mem_size = DeviceInfo.wPageSize;
+       tmp_buf_write_blk_table_data = kmalloc(mem_size, GFP_ATOMIC);
+       if (!tmp_buf_write_blk_table_data)
+               goto tmp_buf_write_blk_table_data_fail;
+       memset(tmp_buf_write_blk_table_data, 0xff, mem_size);
+       total_bytes += mem_size;
+
+       /* Malloc memory for function FTL_Read_Disturbance */
+       tmp_buf_read_disturbance = kmalloc(block_size, GFP_ATOMIC);
+       if (!tmp_buf_read_disturbance)
+               goto tmp_buf_read_disturbance_fail;
+       memset(tmp_buf_read_disturbance, 0xff, block_size);
+       total_bytes += block_size;
+
+       /* Alloc mem for function NAND_Read_Page_Main_Spare of lld_nand.c */
+       buf_read_page_main_spare = kmalloc(DeviceInfo.wPageSize, GFP_ATOMIC);
+       if (!buf_read_page_main_spare)
+               goto buf_read_page_main_spare_fail;
+       total_bytes += DeviceInfo.wPageSize;
+
+       /* Alloc mem for function NAND_Write_Page_Main_Spare of lld_nand.c */
+       buf_write_page_main_spare = kmalloc(DeviceInfo.wPageSize, GFP_ATOMIC);
+       if (!buf_write_page_main_spare)
+               goto buf_write_page_main_spare_fail;
+       total_bytes += DeviceInfo.wPageSize;
+
+       /* Alloc mem for function NAND_Read_Page_Spare of lld_nand.c */
+       buf_read_page_spare = kmalloc(DeviceInfo.wPageSpareSize, GFP_ATOMIC);
+       if (!buf_read_page_spare)
+               goto buf_read_page_spare_fail;
+       memset(buf_read_page_spare, 0xff, DeviceInfo.wPageSpareSize);
+       total_bytes += DeviceInfo.wPageSpareSize;
+
+       /* Alloc mem for function NAND_Get_Bad_Block of lld_nand.c */
+       buf_get_bad_block = kmalloc(DeviceInfo.wPageSpareSize, GFP_ATOMIC);
+       if (!buf_get_bad_block)
+               goto buf_get_bad_block_fail;
+       memset(buf_get_bad_block, 0xff, DeviceInfo.wPageSpareSize);
+       total_bytes += DeviceInfo.wPageSpareSize;
+
+#if CMD_DMA
+       g_temp_buf = kmalloc(block_size, GFP_ATOMIC);
+       if (!g_temp_buf)
+               goto temp_buf_fail;
+       memset(g_temp_buf, 0xff, block_size);
+       total_bytes += block_size;
+
+       /* Malloc memory for copy of block table used in CDMA mode */
+       g_pBTStartingCopy = kmalloc(block_table_size, GFP_ATOMIC);
+       if (!g_pBTStartingCopy)
+               goto bt_starting_copy;
+       memset(g_pBTStartingCopy, 0, block_table_size);
+       total_bytes += block_table_size;
+
+       g_pWearCounterCopy = (u8 *)(g_pBTStartingCopy +
+               DeviceInfo.wDataBlockNum * sizeof(u32));
+
+       if (DeviceInfo.MLCDevice)
+               g_pReadCounterCopy = (u16 *)(g_pBTStartingCopy +
+                       DeviceInfo.wDataBlockNum *
+                       (sizeof(u32) + sizeof(u8)));
+
+       /* Malloc memory for block table copies */
+       mem_size = 5 * DeviceInfo.wDataBlockNum * sizeof(u32) +
+                       5 * DeviceInfo.wDataBlockNum * sizeof(u8);
+       if (DeviceInfo.MLCDevice)
+               mem_size += 5 * DeviceInfo.wDataBlockNum * sizeof(u16);
+       g_pBlockTableCopies = kmalloc(mem_size, GFP_ATOMIC);
+       if (!g_pBlockTableCopies)
+               goto blk_table_copies_fail;
+       memset(g_pBlockTableCopies, 0, mem_size);
+       total_bytes += mem_size;
+       g_pNextBlockTable = g_pBlockTableCopies;
+
+       /* Malloc memory for Block Table Delta */
+       mem_size = MAX_DESCS * sizeof(struct BTableChangesDelta);
+       g_pBTDelta = kmalloc(mem_size, GFP_ATOMIC);
+       if (!g_pBTDelta)
+               goto bt_delta_fail;
+       memset(g_pBTDelta, 0, mem_size);
+       total_bytes += mem_size;
+       g_pBTDelta_Free = g_pBTDelta;
+
+       /* Malloc memory for Copy Back Buffers */
+       for (j = 0; j < COPY_BACK_BUF_NUM; j++) {
+               cp_back_buf_copies[j] = kmalloc(block_size, GFP_ATOMIC);
+               if (!cp_back_buf_copies[j])
+                       goto cp_back_buf_copies_fail;
+               memset(cp_back_buf_copies[j], 0, block_size);
+               total_bytes += block_size;
+       }
+       cp_back_buf_idx = 0;
+
+       /* Malloc memory for pending commands list */
+       mem_size = sizeof(struct pending_cmd) * MAX_DESCS;
+       info.pcmds = kzalloc(mem_size, GFP_KERNEL);
+       if (!info.pcmds)
+               goto pending_cmds_buf_fail;
+       total_bytes += mem_size;
+
+       /* Malloc memory for CDMA descripter table */
+       mem_size = sizeof(struct cdma_descriptor) * MAX_DESCS;
+       info.cdma_desc_buf = kzalloc(mem_size, GFP_KERNEL);
+       if (!info.cdma_desc_buf)
+               goto cdma_desc_buf_fail;
+       total_bytes += mem_size;
+
+       /* Malloc memory for Memcpy descripter table */
+       mem_size = sizeof(struct memcpy_descriptor) * MAX_DESCS;
+       info.memcp_desc_buf = kzalloc(mem_size, GFP_KERNEL);
+       if (!info.memcp_desc_buf)
+               goto memcp_desc_buf_fail;
+       total_bytes += mem_size;
+#endif
+
+       nand_dbg_print(NAND_DBG_WARN,
+               "Total memory allocated in FTL layer: %d\n", total_bytes);
+
+       return PASS;
+
+#if CMD_DMA
+memcp_desc_buf_fail:
+       kfree(info.cdma_desc_buf);
+cdma_desc_buf_fail:
+       kfree(info.pcmds);
+pending_cmds_buf_fail:
+cp_back_buf_copies_fail:
+       j--;
+       for (; j >= 0; j--)
+               kfree(cp_back_buf_copies[j]);
+       kfree(g_pBTDelta);
+bt_delta_fail:
+       kfree(g_pBlockTableCopies);
+blk_table_copies_fail:
+       kfree(g_pBTStartingCopy);
+bt_starting_copy:
+       kfree(g_temp_buf);
+temp_buf_fail:
+       kfree(buf_get_bad_block);
+#endif
+
+buf_get_bad_block_fail:
+       kfree(buf_read_page_spare);
+buf_read_page_spare_fail:
+       kfree(buf_write_page_main_spare);
+buf_write_page_main_spare_fail:
+       kfree(buf_read_page_main_spare);
+buf_read_page_main_spare_fail:
+       kfree(tmp_buf_read_disturbance);
+tmp_buf_read_disturbance_fail:
+       kfree(tmp_buf_write_blk_table_data);
+tmp_buf_write_blk_table_data_fail:
+       kfree(flags_static_wear_leveling);
+flags_static_wear_leveling_fail:
+       kfree(tmp_buf2_read_blk_table);
+tmp_buf2_read_blk_table_fail:
+       kfree(tmp_buf1_read_blk_table);
+tmp_buf1_read_blk_table_fail:
+       kfree(spare_buf_bt_search_bt_in_block);
+spare_buf_bt_search_bt_in_block_fail:
+       kfree(spare_buf_search_bt_in_block);
+spare_buf_search_bt_in_block_fail:
+       kfree(tmp_buf_search_bt_in_block);
+tmp_buf_search_bt_in_block_fail:
+       kfree(flag_check_blk_table);
+flag_check_blk_table_fail:
+       kfree(g_pBTBlocks);
+bt_blocks_fail:
+       kfree(g_pTempBuf);
+Temp_buf_fail:
+       kfree(cache_l2_blk_buf);
+cache_l2_blk_buf_fail:
+       kfree(cache_l2_page_buf);
+cache_l2_page_buf_fail:
+       kfree(g_pIPF);
+ipf_fail:
+cache_item_fail:
+       i--;
+       for (; i >= 0; i--)
+               kfree(Cache.array[i].buf);
+       kfree(g_pBlockTable);
+block_table_fail:
+       printk(KERN_ERR "Failed to kmalloc memory in %s Line %d.\n",
+               __FILE__, __LINE__);
+
+       return -ENOMEM;
+}
+
+/* .... */
+static int free_memory(void)
+{
+       int i;
+
+#if CMD_DMA
+       kfree(info.memcp_desc_buf);
+       kfree(info.cdma_desc_buf);
+       kfree(info.pcmds);
+       for (i = COPY_BACK_BUF_NUM - 1; i >= 0; i--)
+               kfree(cp_back_buf_copies[i]);
+       kfree(g_pBTDelta);
+       kfree(g_pBlockTableCopies);
+       kfree(g_pBTStartingCopy);
+       kfree(g_temp_buf);
+       kfree(buf_get_bad_block);
+#endif
+       kfree(buf_read_page_spare);
+       kfree(buf_write_page_main_spare);
+       kfree(buf_read_page_main_spare);
+       kfree(tmp_buf_read_disturbance);
+       kfree(tmp_buf_write_blk_table_data);
+       kfree(flags_static_wear_leveling);
+       kfree(tmp_buf2_read_blk_table);
+       kfree(tmp_buf1_read_blk_table);
+       kfree(spare_buf_bt_search_bt_in_block);
+       kfree(spare_buf_search_bt_in_block);
+       kfree(tmp_buf_search_bt_in_block);
+       kfree(flag_check_blk_table);
+       kfree(g_pBTBlocks);
+       kfree(g_pTempBuf);
+       kfree(g_pIPF);
+       for (i = CACHE_ITEM_NUM - 1; i >= 0; i--)
+               kfree(Cache.array[i].buf);
+       kfree(g_pBlockTable);
+
+       return 0;
+}
+
+static void dump_cache_l2_table(void)
+{
+       struct list_head *p;
+       struct spectra_l2_cache_list *pnd;
+       int n, i;
+
+       n = 0;
+       list_for_each(p, &cache_l2.table.list) {
+               pnd = list_entry(p, struct spectra_l2_cache_list, list);
+               nand_dbg_print(NAND_DBG_WARN, "dump_cache_l2_table node: %d, logical_blk_num: %d\n", n, pnd->logical_blk_num);
+/*
+               for (i = 0; i < DeviceInfo.wPagesPerBlock; i++) {
+                       if (pnd->pages_array[i] != MAX_U32_VALUE)
+                               nand_dbg_print(NAND_DBG_WARN, "    pages_array[%d]: 0x%x\n", i, pnd->pages_array[i]);
+               }
+*/
+               n++;
+       }
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Init
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=1
+* Description:  allocates the memory for cache array,
+*               important data structures
+*               clears the cache array
+*               reads the block table from flash into array
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Init(void)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       Cache.pages_per_item = 1;
+       Cache.cache_item_size = 1 * DeviceInfo.wPageDataSize;
+
+       if (allocate_memory() != PASS)
+               return FAIL;
+
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       memcpy((void *)&cache_start_copy, (void *)&Cache,
+               sizeof(struct flash_cache_tag));
+       memset((void *)&int_cache, -1,
+               sizeof(struct flash_cache_delta_list_tag) *
+               (MAX_CHANS + MAX_DESCS));
+#endif
+       ftl_cmd_cnt = 0;
+#endif
+
+       if (FTL_Read_Block_Table() != PASS)
+               return FAIL;
+
+       /* Init the Level2 Cache data structure */
+       for (i = 0; i < BLK_NUM_FOR_L2_CACHE; i++)
+               cache_l2.blk_array[i] = MAX_U32_VALUE;
+       cache_l2.cur_blk_idx = 0;
+       cache_l2.cur_page_num = 0;
+       INIT_LIST_HEAD(&cache_l2.table.list);
+       cache_l2.table.logical_blk_num = MAX_U32_VALUE;
+
+       dump_cache_l2_table();
+
+       return 0;
+}
+
+
+#if CMD_DMA
+#if 0
+static void save_blk_table_changes(u16 idx)
+{
+       u8 ftl_cmd;
+       u32 *pbt = (u32 *)g_pBTStartingCopy;
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       u16 id;
+       u8 cache_blks;
+
+       id = idx - MAX_CHANS;
+       if (int_cache[id].item != -1) {
+               cache_blks = int_cache[id].item;
+               cache_start_copy.array[cache_blks].address =
+                       int_cache[id].cache.address;
+               cache_start_copy.array[cache_blks].changed =
+                       int_cache[id].cache.changed;
+       }
+#endif
+
+       ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+
+       while (ftl_cmd <= PendingCMD[idx].Tag) {
+               if (p_BTableChangesDelta->ValidFields == 0x01) {
+                       g_wBlockTableOffset =
+                               p_BTableChangesDelta->g_wBlockTableOffset;
+               } else if (p_BTableChangesDelta->ValidFields == 0x0C) {
+                       pbt[p_BTableChangesDelta->BT_Index] =
+                               p_BTableChangesDelta->BT_Entry_Value;
+                       debug_boundary_error(((
+                               p_BTableChangesDelta->BT_Index)),
+                               DeviceInfo.wDataBlockNum, 0);
+               } else if (p_BTableChangesDelta->ValidFields == 0x03) {
+                       g_wBlockTableOffset =
+                               p_BTableChangesDelta->g_wBlockTableOffset;
+                       g_wBlockTableIndex =
+                               p_BTableChangesDelta->g_wBlockTableIndex;
+               } else if (p_BTableChangesDelta->ValidFields == 0x30) {
+                       g_pWearCounterCopy[p_BTableChangesDelta->WC_Index] =
+                               p_BTableChangesDelta->WC_Entry_Value;
+               } else if ((DeviceInfo.MLCDevice) &&
+                       (p_BTableChangesDelta->ValidFields == 0xC0)) {
+                       g_pReadCounterCopy[p_BTableChangesDelta->RC_Index] =
+                               p_BTableChangesDelta->RC_Entry_Value;
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "In event status setting read counter "
+                               "GLOB_ftl_cmd_cnt %u Count %u Index %u\n",
+                               ftl_cmd,
+                               p_BTableChangesDelta->RC_Entry_Value,
+                               (unsigned int)p_BTableChangesDelta->RC_Index);
+               } else {
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "This should never occur \n");
+               }
+               p_BTableChangesDelta += 1;
+               ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+       }
+}
+
+static void discard_cmds(u16 n)
+{
+       u32 *pbt = (u32 *)g_pBTStartingCopy;
+       u8 ftl_cmd;
+       unsigned long k;
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       u8 cache_blks;
+       u16 id;
+#endif
+
+       if ((PendingCMD[n].CMD == WRITE_MAIN_CMD) ||
+               (PendingCMD[n].CMD == WRITE_MAIN_SPARE_CMD)) {
+               for (k = 0; k < DeviceInfo.wDataBlockNum; k++) {
+                       if (PendingCMD[n].Block == (pbt[k] & (~BAD_BLOCK)))
+                               MARK_BLK_AS_DISCARD(pbt[k]);
+               }
+       }
+
+       ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+       while (ftl_cmd <= PendingCMD[n].Tag) {
+               p_BTableChangesDelta += 1;
+               ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+       }
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       id = n - MAX_CHANS;
+
+       if (int_cache[id].item != -1) {
+               cache_blks = int_cache[id].item;
+               if (PendingCMD[n].CMD == MEMCOPY_CMD) {
+                       if ((cache_start_copy.array[cache_blks].buf <=
+                               PendingCMD[n].DataDestAddr) &&
+                               ((cache_start_copy.array[cache_blks].buf +
+                               Cache.cache_item_size) >
+                               PendingCMD[n].DataDestAddr)) {
+                               cache_start_copy.array[cache_blks].address =
+                                               NAND_CACHE_INIT_ADDR;
+                               cache_start_copy.array[cache_blks].use_cnt =
+                                                               0;
+                               cache_start_copy.array[cache_blks].changed =
+                                                               CLEAR;
+                       }
+               } else {
+                       cache_start_copy.array[cache_blks].address =
+                                       int_cache[id].cache.address;
+                       cache_start_copy.array[cache_blks].changed =
+                                       int_cache[id].cache.changed;
+               }
+       }
+#endif
+}
+
+static void process_cmd_pass(int *first_failed_cmd, u16 idx)
+{
+       if (0 == *first_failed_cmd)
+               save_blk_table_changes(idx);
+       else
+               discard_cmds(idx);
+}
+
+static void process_cmd_fail_abort(int *first_failed_cmd,
+                               u16 idx, int event)
+{
+       u32 *pbt = (u32 *)g_pBTStartingCopy;
+       u8 ftl_cmd;
+       unsigned long i;
+       int erase_fail, program_fail;
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       u8 cache_blks;
+       u16 id;
+#endif
+
+       if (0 == *first_failed_cmd)
+               *first_failed_cmd = PendingCMD[idx].SBDCmdIndex;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Uncorrectable error has occured "
+               "while executing %u Command %u accesing Block %u\n",
+               (unsigned int)p_BTableChangesDelta->ftl_cmd_cnt,
+               PendingCMD[idx].CMD,
+               (unsigned int)PendingCMD[idx].Block);
+
+       ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+       while (ftl_cmd <= PendingCMD[idx].Tag) {
+               p_BTableChangesDelta += 1;
+               ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+       }
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       id = idx - MAX_CHANS;
+
+       if (int_cache[id].item != -1) {
+               cache_blks = int_cache[id].item;
+               if ((PendingCMD[idx].CMD == WRITE_MAIN_CMD)) {
+                       cache_start_copy.array[cache_blks].address =
+                                       int_cache[id].cache.address;
+                       cache_start_copy.array[cache_blks].changed = SET;
+               } else if ((PendingCMD[idx].CMD == READ_MAIN_CMD)) {
+                       cache_start_copy.array[cache_blks].address =
+                               NAND_CACHE_INIT_ADDR;
+                       cache_start_copy.array[cache_blks].use_cnt = 0;
+                       cache_start_copy.array[cache_blks].changed =
+                                                       CLEAR;
+               } else if (PendingCMD[idx].CMD == ERASE_CMD) {
+                       /* ? */
+               } else if (PendingCMD[idx].CMD == MEMCOPY_CMD) {
+                       /* ? */
+               }
+       }
+#endif
+
+       erase_fail = (event == EVENT_ERASE_FAILURE) &&
+                       (PendingCMD[idx].CMD == ERASE_CMD);
+
+       program_fail = (event == EVENT_PROGRAM_FAILURE) &&
+                       ((PendingCMD[idx].CMD == WRITE_MAIN_CMD) ||
+                       (PendingCMD[idx].CMD == WRITE_MAIN_SPARE_CMD));
+
+       if (erase_fail || program_fail) {
+               for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+                       if (PendingCMD[idx].Block ==
+                               (pbt[i] & (~BAD_BLOCK)))
+                               MARK_BLOCK_AS_BAD(pbt[i]);
+               }
+       }
+}
+
+static void process_cmd(int *first_failed_cmd, u16 idx, int event)
+{
+       u8 ftl_cmd;
+       int cmd_match = 0;
+
+       if (p_BTableChangesDelta->ftl_cmd_cnt == PendingCMD[idx].Tag)
+               cmd_match = 1;
+
+       if (PendingCMD[idx].Status == CMD_PASS) {
+               process_cmd_pass(first_failed_cmd, idx);
+       } else if ((PendingCMD[idx].Status == CMD_FAIL) ||
+                       (PendingCMD[idx].Status == CMD_ABORT)) {
+               process_cmd_fail_abort(first_failed_cmd, idx, event);
+       } else if ((PendingCMD[idx].Status == CMD_NOT_DONE) &&
+                                       PendingCMD[idx].Tag) {
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       " Command no. %hu is not executed\n",
+                       (unsigned int)PendingCMD[idx].Tag);
+               ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+               while (ftl_cmd <= PendingCMD[idx].Tag) {
+                       p_BTableChangesDelta += 1;
+                       ftl_cmd = p_BTableChangesDelta->ftl_cmd_cnt;
+               }
+       }
+}
+#endif
+
+static void process_cmd(int *first_failed_cmd, u16 idx, int event)
+{
+       printk(KERN_ERR "temporary workaround function. "
+               "Should not be called! \n");
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:            GLOB_FTL_Event_Status
+* Inputs:       none
+* Outputs:      Event Code
+* Description: It is called by SBD after hardware interrupt signalling
+*               completion of commands chain
+*               It does following things
+*               get event status from LLD
+*               analyze command chain status
+*               determine last command executed
+*               analyze results
+*               rebuild the block table in case of uncorrectable error
+*               return event code
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Event_Status(int *first_failed_cmd)
+{
+       int event_code = PASS;
+       u16 i_P;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       *first_failed_cmd = 0;
+
+       event_code = GLOB_LLD_Event_Status();
+
+       switch (event_code) {
+       case EVENT_PASS:
+               nand_dbg_print(NAND_DBG_DEBUG, "Handling EVENT_PASS\n");
+               break;
+       case EVENT_UNCORRECTABLE_DATA_ERROR:
+               nand_dbg_print(NAND_DBG_DEBUG, "Handling Uncorrectable ECC!\n");
+               break;
+       case EVENT_PROGRAM_FAILURE:
+       case EVENT_ERASE_FAILURE:
+               nand_dbg_print(NAND_DBG_WARN, "Handling Ugly case. "
+                       "Event code: 0x%x\n", event_code);
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta;
+               for (i_P = MAX_CHANS; i_P < (ftl_cmd_cnt + MAX_CHANS);
+                               i_P++)
+                       process_cmd(first_failed_cmd, i_P, event_code);
+               memcpy(g_pBlockTable, g_pBTStartingCopy,
+                       DeviceInfo.wDataBlockNum * sizeof(u32));
+               memcpy(g_pWearCounter, g_pWearCounterCopy,
+                       DeviceInfo.wDataBlockNum * sizeof(u8));
+               if (DeviceInfo.MLCDevice)
+                       memcpy(g_pReadCounter, g_pReadCounterCopy,
+                               DeviceInfo.wDataBlockNum * sizeof(u16));
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+               memcpy((void *)&Cache, (void *)&cache_start_copy,
+                       sizeof(struct flash_cache_tag));
+               memset((void *)&int_cache, -1,
+                       sizeof(struct flash_cache_delta_list_tag) *
+                       (MAX_DESCS + MAX_CHANS));
+#endif
+               break;
+       default:
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Handling unexpected event code - 0x%x\n",
+                       event_code);
+               event_code = ERR;
+               break;
+       }
+
+       memcpy(g_pBTStartingCopy, g_pBlockTable,
+               DeviceInfo.wDataBlockNum * sizeof(u32));
+       memcpy(g_pWearCounterCopy, g_pWearCounter,
+               DeviceInfo.wDataBlockNum * sizeof(u8));
+       if (DeviceInfo.MLCDevice)
+               memcpy(g_pReadCounterCopy, g_pReadCounter,
+                       DeviceInfo.wDataBlockNum * sizeof(u16));
+
+       g_pBTDelta_Free = g_pBTDelta;
+       ftl_cmd_cnt = 0;
+       g_pNextBlockTable = g_pBlockTableCopies;
+       cp_back_buf_idx = 0;
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       memcpy((void *)&cache_start_copy, (void *)&Cache,
+               sizeof(struct flash_cache_tag));
+       memset((void *)&int_cache, -1,
+               sizeof(struct flash_cache_delta_list_tag) *
+               (MAX_DESCS + MAX_CHANS));
+#endif
+
+       return event_code;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     glob_ftl_execute_cmds
+* Inputs:       none
+* Outputs:      none
+* Description:  pass thru to LLD
+***************************************************************/
+u16 glob_ftl_execute_cmds(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE,
+               "glob_ftl_execute_cmds: ftl_cmd_cnt %u\n",
+               (unsigned int)ftl_cmd_cnt);
+       g_SBDCmdIndex = 0;
+       return glob_lld_execute_cmds();
+}
+
+#endif
+
+#if !CMD_DMA
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Read Immediate
+* Inputs:         pointer to data
+*                     address of data
+* Outputs:      PASS / FAIL
+* Description:  Reads one page of data into RAM directly from flash without
+*       using or disturbing cache.It is assumed this function is called
+*       with CMD-DMA disabled.
+*****************************************************************/
+int GLOB_FTL_Read_Immediate(u8 *read_data, u64 addr)
+{
+       int wResult = FAIL;
+       u32 Block;
+       u16 Page;
+       u32 phy_blk;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       Block = BLK_FROM_ADDR(addr);
+       Page = PAGE_FROM_ADDR(addr, Block);
+
+       if (!IS_SPARE_BLOCK(Block))
+               return FAIL;
+
+       phy_blk = pbt[Block];
+       wResult = GLOB_LLD_Read_Page_Main(read_data, phy_blk, Page, 1);
+
+       if (DeviceInfo.MLCDevice) {
+               g_pReadCounter[phy_blk - DeviceInfo.wSpectraStartBlock]++;
+               if (g_pReadCounter[phy_blk - DeviceInfo.wSpectraStartBlock]
+                       >= MAX_READ_COUNTER)
+                       FTL_Read_Disturbance(phy_blk);
+               if (g_cBlockTableStatus != IN_PROGRESS_BLOCK_TABLE) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+       }
+
+       return wResult;
+}
+#endif
+
+#ifdef SUPPORT_BIG_ENDIAN
+/*********************************************************************
+* Function:     FTL_Invert_Block_Table
+* Inputs:       none
+* Outputs:      none
+* Description:  Re-format the block table in ram based on BIG_ENDIAN and
+*                     LARGE_BLOCKNUM if necessary
+**********************************************************************/
+static void FTL_Invert_Block_Table(void)
+{
+       u32 i;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+#ifdef SUPPORT_LARGE_BLOCKNUM
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               pbt[i] = INVERTUINT32(pbt[i]);
+               g_pWearCounter[i] = INVERTUINT32(g_pWearCounter[i]);
+       }
+#else
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               pbt[i] = INVERTUINT16(pbt[i]);
+               g_pWearCounter[i] = INVERTUINT16(g_pWearCounter[i]);
+       }
+#endif
+}
+#endif
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Flash_Init
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=0x01 (based on read ID)
+* Description:  The flash controller is initialized
+*               The flash device is reset
+*               Perform a flash READ ID command to confirm that a
+*                   valid device is attached and active.
+*                   The DeviceInfo structure gets filled in
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Flash_Init(void)
+{
+       int status = FAIL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       g_SBDCmdIndex = 0;
+
+       GLOB_LLD_Flash_Init();
+
+       status = GLOB_LLD_Read_Device_ID();
+
+       return status;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=0x01 (based on read ID)
+* Description:  The flash controller is released
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Flash_Release(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       return GLOB_LLD_Flash_Release();
+}
+
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Cache_Release
+* Inputs:       none
+* Outputs:      none
+* Description:  release all allocated memory in GLOB_FTL_Init
+*               (allocated in GLOB_FTL_Init)
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+void GLOB_FTL_Cache_Release(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       free_memory();
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_If_Hit
+* Inputs:       Page Address
+* Outputs:      Block number/UNHIT BLOCK
+* Description:  Determines if the addressed page is in cache
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u16 FTL_Cache_If_Hit(u64 page_addr)
+{
+       u16 item;
+       u64 addr;
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       item = UNHIT_CACHE_ITEM;
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               addr = Cache.array[i].address;
+               if ((page_addr >= addr) &&
+                       (page_addr < (addr + Cache.cache_item_size))) {
+                       item = i;
+                       break;
+               }
+       }
+
+       return item;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Calculate_LRU
+* Inputs:       None
+* Outputs:      None
+* Description:  Calculate the least recently block in a cache and record its
+*               index in LRU field.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static void FTL_Calculate_LRU(void)
+{
+       u16 i, bCurrentLRU, bTempCount;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       bCurrentLRU = 0;
+       bTempCount = MAX_WORD_VALUE;
+
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               if (Cache.array[i].use_cnt < bTempCount) {
+                       bCurrentLRU = i;
+                       bTempCount = Cache.array[i].use_cnt;
+               }
+       }
+
+       Cache.LRU = bCurrentLRU;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Read_Page
+* Inputs:       pointer to read buffer, logical address and cache item number
+* Outputs:      None
+* Description:  Read the page from the cached block addressed by blocknumber
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static void FTL_Cache_Read_Page(u8 *data_buf, u64 logic_addr, u16 cache_item)
+{
+       u8 *start_addr;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       start_addr = Cache.array[cache_item].buf;
+       start_addr += (u32)(((logic_addr - Cache.array[cache_item].address) >>
+               DeviceInfo.nBitsInPageDataSize) * DeviceInfo.wPageDataSize);
+
+#if CMD_DMA
+       GLOB_LLD_MemCopy_CMD(data_buf, start_addr,
+                       DeviceInfo.wPageDataSize, 0);
+       ftl_cmd_cnt++;
+#else
+       memcpy(data_buf, start_addr, DeviceInfo.wPageDataSize);
+#endif
+
+       if (Cache.array[cache_item].use_cnt < MAX_WORD_VALUE)
+               Cache.array[cache_item].use_cnt++;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Read_All
+* Inputs:       pointer to read buffer,block address
+* Outputs:      PASS=0 / FAIL =1
+* Description:  It reads pages in cache
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Read_All(u8 *pData, u64 phy_addr)
+{
+       int wResult = PASS;
+       u32 Block;
+       u32 lba;
+       u16 Page;
+       u16 PageCount;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 i;
+
+       Block = BLK_FROM_ADDR(phy_addr);
+       Page = PAGE_FROM_ADDR(phy_addr, Block);
+       PageCount = Cache.pages_per_item;
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+                       "%s, Line %d, Function: %s, Block: 0x%x\n",
+                       __FILE__, __LINE__, __func__, Block);
+
+       lba = 0xffffffff;
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if ((pbt[i] & (~BAD_BLOCK)) == Block) {
+                       lba = i;
+                       if (IS_SPARE_BLOCK(i) || IS_BAD_BLOCK(i) ||
+                               IS_DISCARDED_BLOCK(i)) {
+                               /* Add by yunpeng -2008.12.3 */
+#if CMD_DMA
+                               GLOB_LLD_MemCopy_CMD(pData, g_temp_buf,
+                               PageCount * DeviceInfo.wPageDataSize, 0);
+                               ftl_cmd_cnt++;
+#else
+                               memset(pData, 0xFF,
+                                       PageCount * DeviceInfo.wPageDataSize);
+#endif
+                               return wResult;
+                       } else {
+                               continue; /* break ?? */
+                       }
+               }
+       }
+
+       if (0xffffffff == lba)
+               printk(KERN_ERR "FTL_Cache_Read_All: Block is not found in BT\n");
+
+#if CMD_DMA
+       wResult = GLOB_LLD_Read_Page_Main_cdma(pData, Block, Page,
+                       PageCount, LLD_CMD_FLAG_MODE_CDMA);
+       if (DeviceInfo.MLCDevice) {
+               g_pReadCounter[Block - DeviceInfo.wSpectraStartBlock]++;
+               nand_dbg_print(NAND_DBG_DEBUG,
+                              "Read Counter modified in ftl_cmd_cnt %u"
+                               " Block %u Counter%u\n",
+                              ftl_cmd_cnt, (unsigned int)Block,
+                              g_pReadCounter[Block -
+                              DeviceInfo.wSpectraStartBlock]);
+
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+               p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+               p_BTableChangesDelta->RC_Index =
+                       Block - DeviceInfo.wSpectraStartBlock;
+               p_BTableChangesDelta->RC_Entry_Value =
+                       g_pReadCounter[Block - DeviceInfo.wSpectraStartBlock];
+               p_BTableChangesDelta->ValidFields = 0xC0;
+
+               ftl_cmd_cnt++;
+
+               if (g_pReadCounter[Block - DeviceInfo.wSpectraStartBlock] >=
+                   MAX_READ_COUNTER)
+                       FTL_Read_Disturbance(Block);
+               if (g_cBlockTableStatus != IN_PROGRESS_BLOCK_TABLE) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+       } else {
+               ftl_cmd_cnt++;
+       }
+#else
+       wResult = GLOB_LLD_Read_Page_Main(pData, Block, Page, PageCount);
+       if (wResult == FAIL)
+               return wResult;
+
+       if (DeviceInfo.MLCDevice) {
+               g_pReadCounter[Block - DeviceInfo.wSpectraStartBlock]++;
+               if (g_pReadCounter[Block - DeviceInfo.wSpectraStartBlock] >=
+                                               MAX_READ_COUNTER)
+                       FTL_Read_Disturbance(Block);
+               if (g_cBlockTableStatus != IN_PROGRESS_BLOCK_TABLE) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+       }
+#endif
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Write_All
+* Inputs:       pointer to cache in sys memory
+*               address of free block in flash
+* Outputs:      PASS=0 / FAIL=1
+* Description:  writes all the pages of the block in cache to flash
+*
+*               NOTE:need to make sure this works ok when cache is limited
+*               to a partial block. This is where copy-back would be
+*               activated.  This would require knowing which pages in the
+*               cached block are clean/dirty.Right now we only know if
+*               the whole block is clean/dirty.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Write_All(u8 *pData, u64 blk_addr)
+{
+       u16 wResult = PASS;
+       u32 Block;
+       u16 Page;
+       u16 PageCount;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       nand_dbg_print(NAND_DBG_DEBUG, "This block %d going to be written "
+               "on %d\n", cache_block_to_write,
+               (u32)(blk_addr >> DeviceInfo.nBitsInBlockDataSize));
+
+       Block = BLK_FROM_ADDR(blk_addr);
+       Page = PAGE_FROM_ADDR(blk_addr, Block);
+       PageCount = Cache.pages_per_item;
+
+#if CMD_DMA
+       if (FAIL == GLOB_LLD_Write_Page_Main_cdma(pData,
+                                       Block, Page, PageCount)) {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "NAND Program fail in %s, Line %d, "
+                       "Function: %s, new Bad Block %d generated! "
+                       "Need Bad Block replacing.\n",
+                       __FILE__, __LINE__, __func__, Block);
+               wResult = FAIL;
+       }
+       ftl_cmd_cnt++;
+#else
+       if (FAIL == GLOB_LLD_Write_Page_Main(pData, Block, Page, PageCount)) {
+               nand_dbg_print(NAND_DBG_WARN, "NAND Program fail in %s,"
+                       " Line %d, Function %s, new Bad Block %d generated!"
+                       "Need Bad Block replacing.\n",
+                       __FILE__, __LINE__, __func__, Block);
+               wResult = FAIL;
+       }
+#endif
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Update_Block
+* Inputs:       pointer to buffer,page address,block address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It updates the cache
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Update_Block(u8 *pData,
+                       u64 old_page_addr, u64 blk_addr)
+{
+       int i, j;
+       u8 *buf = pData;
+       int wResult = PASS;
+       int wFoundInCache;
+       u64 page_addr;
+       u64 addr;
+       u64 old_blk_addr;
+       u16 page_offset;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                               __FILE__, __LINE__, __func__);
+
+       old_blk_addr = (u64)(old_page_addr >>
+               DeviceInfo.nBitsInBlockDataSize) * DeviceInfo.wBlockDataSize;
+       page_offset = (u16)(GLOB_u64_Remainder(old_page_addr, 2) >>
+               DeviceInfo.nBitsInPageDataSize);
+
+       for (i = 0; i < DeviceInfo.wPagesPerBlock; i += Cache.pages_per_item) {
+               page_addr = old_blk_addr + i * DeviceInfo.wPageDataSize;
+               if (i != page_offset) {
+                       wFoundInCache = FAIL;
+                       for (j = 0; j < CACHE_ITEM_NUM; j++) {
+                               addr = Cache.array[j].address;
+                               addr = FTL_Get_Physical_Block_Addr(addr) +
+                                       GLOB_u64_Remainder(addr, 2);
+                               if ((addr >= page_addr) && addr <
+                                       (page_addr + Cache.cache_item_size)) {
+                                       wFoundInCache = PASS;
+                                       buf = Cache.array[j].buf;
+                                       Cache.array[j].changed = SET;
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+                                       int_cache[ftl_cmd_cnt].item = j;
+                                       int_cache[ftl_cmd_cnt].cache.address =
+                                               Cache.array[j].address;
+                                       int_cache[ftl_cmd_cnt].cache.changed =
+                                               Cache.array[j].changed;
+#endif
+#endif
+                                       break;
+                               }
+                       }
+                       if (FAIL == wFoundInCache) {
+                               if (ERR == FTL_Cache_Read_All(g_pTempBuf,
+                                       page_addr)) {
+                                       wResult = FAIL;
+                                       break;
+                               }
+                               buf = g_pTempBuf;
+                       }
+               } else {
+                       buf = pData;
+               }
+
+               if (FAIL == FTL_Cache_Write_All(buf,
+                       blk_addr + (page_addr - old_blk_addr))) {
+                       wResult = FAIL;
+                       break;
+               }
+       }
+
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Copy_Block
+* Inputs:       source block address
+*               Destination block address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  used only for static wear leveling to move the block
+*               containing static data to new blocks(more worn)
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int FTL_Copy_Block(u64 old_blk_addr, u64 blk_addr)
+{
+       int i, r1, r2, wResult = PASS;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < DeviceInfo.wPagesPerBlock; i += Cache.pages_per_item) {
+               r1 = FTL_Cache_Read_All(g_pTempBuf, old_blk_addr +
+                                       i * DeviceInfo.wPageDataSize);
+               r2 = FTL_Cache_Write_All(g_pTempBuf, blk_addr +
+                                       i * DeviceInfo.wPageDataSize);
+               if ((ERR == r1) || (FAIL == r2)) {
+                       wResult = FAIL;
+                       break;
+               }
+       }
+
+       return wResult;
+}
+
+/* Search the block table to find out the least wear block and then return it */
+static u32 find_least_worn_blk_for_l2_cache(void)
+{
+       int i;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u8 least_wear_cnt = MAX_BYTE_VALUE;
+       u32 least_wear_blk_idx = MAX_U32_VALUE;
+       u32 phy_idx;
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_SPARE_BLOCK(i)) {
+                       phy_idx = (u32)((~BAD_BLOCK) & pbt[i]);
+                       if (phy_idx > DeviceInfo.wSpectraEndBlock)
+                               printk(KERN_ERR "find_least_worn_blk_for_l2_cache: "
+                                       "Too big phy block num (%d)\n", phy_idx);
+                       if (g_pWearCounter[phy_idx -DeviceInfo.wSpectraStartBlock] < least_wear_cnt) {
+                               least_wear_cnt = g_pWearCounter[phy_idx - DeviceInfo.wSpectraStartBlock];
+                               least_wear_blk_idx = i;
+                       }
+               }
+       }
+
+       nand_dbg_print(NAND_DBG_WARN,
+               "find_least_worn_blk_for_l2_cache: "
+               "find block %d with least worn counter (%d)\n",
+               least_wear_blk_idx, least_wear_cnt);
+
+       return least_wear_blk_idx;
+}
+
+
+
+/* Get blocks for Level2 Cache */
+static int get_l2_cache_blks(void)
+{
+       int n;
+       u32 blk;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       for (n = 0; n < BLK_NUM_FOR_L2_CACHE; n++) {
+               blk = find_least_worn_blk_for_l2_cache();
+               if (blk > DeviceInfo.wDataBlockNum) {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "find_least_worn_blk_for_l2_cache: "
+                               "No enough free NAND blocks (n: %d) for L2 Cache!\n", n);
+                       return FAIL;
+               }
+               /* Tag the free block as discard in block table */
+               pbt[blk] = (pbt[blk] & (~BAD_BLOCK)) | DISCARD_BLOCK;
+               /* Add the free block to the L2 Cache block array */
+               cache_l2.blk_array[n] = pbt[blk] & (~BAD_BLOCK);
+       }
+
+       return PASS;
+}
+
+static int erase_l2_cache_blocks(void)
+{
+       int i, ret = PASS;
+       u32 pblk, lblk;
+       u64 addr;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < BLK_NUM_FOR_L2_CACHE; i++) {
+               pblk = cache_l2.blk_array[i];
+
+               /* If the L2 cache block is invalid, then just skip it */
+               if (MAX_U32_VALUE == pblk)
+                       continue;
+
+               BUG_ON(pblk > DeviceInfo.wSpectraEndBlock);
+
+               addr = (u64)pblk << DeviceInfo.nBitsInBlockDataSize;
+               if (PASS == GLOB_FTL_Block_Erase(addr)) {
+                       /* Get logical block number of the erased block */
+                       lblk = FTL_Get_Block_Index(pblk);
+                       BUG_ON(BAD_BLOCK == lblk);
+                       /* Tag it as free in the block table */
+                       pbt[lblk] &= (u32)(~DISCARD_BLOCK);
+                       pbt[lblk] |= (u32)(SPARE_BLOCK);
+               } else {
+                       MARK_BLOCK_AS_BAD(pbt[lblk]);
+                       ret = ERR;
+               }
+       }
+
+       return ret;
+}
+
+/*
+ * Merge the valid data page in the L2 cache blocks into NAND.
+*/
+static int flush_l2_cache(void)
+{
+       struct list_head *p;
+       struct spectra_l2_cache_list *pnd, *tmp_pnd;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 phy_blk, l2_blk;
+       u64 addr;
+       u16 l2_page;
+       int i, ret = PASS;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       if (list_empty(&cache_l2.table.list)) /* No data to flush */
+               return ret;
+
+       //dump_cache_l2_table();
+
+       if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+               g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+               FTL_Write_IN_Progress_Block_Table_Page();
+       }
+
+       list_for_each(p, &cache_l2.table.list) {
+               pnd = list_entry(p, struct spectra_l2_cache_list, list);
+               if (IS_SPARE_BLOCK(pnd->logical_blk_num) ||
+                       IS_BAD_BLOCK(pnd->logical_blk_num) ||
+                       IS_DISCARDED_BLOCK(pnd->logical_blk_num)) {
+                       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d\n", __FILE__, __LINE__);
+                       memset(cache_l2_blk_buf, 0xff, DeviceInfo.wPagesPerBlock * DeviceInfo.wPageDataSize);                   
+               } else {
+                       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d\n", __FILE__, __LINE__);
+                       phy_blk = pbt[pnd->logical_blk_num] & (~BAD_BLOCK);
+                       ret = GLOB_LLD_Read_Page_Main(cache_l2_blk_buf,
+                               phy_blk, 0, DeviceInfo.wPagesPerBlock);
+                       if (ret == FAIL) {
+                               printk(KERN_ERR "Read NAND page fail in %s, Line %d\n", __FILE__, __LINE__);
+                       }
+               }
+
+               for (i = 0; i < DeviceInfo.wPagesPerBlock; i++) {
+                       if (pnd->pages_array[i] != MAX_U32_VALUE) {
+                               l2_blk = cache_l2.blk_array[(pnd->pages_array[i] >> 16) & 0xffff];
+                               l2_page = pnd->pages_array[i] & 0xffff;
+                               ret = GLOB_LLD_Read_Page_Main(cache_l2_page_buf, l2_blk, l2_page, 1);
+                               if (ret == FAIL) {
+                                       printk(KERN_ERR "Read NAND page fail in %s, Line %d\n", __FILE__, __LINE__);
+                               }
+                               memcpy(cache_l2_blk_buf + i * DeviceInfo.wPageDataSize, cache_l2_page_buf, DeviceInfo.wPageDataSize);
+                       }
+               }
+
+               /* Find a free block and tag the original block as discarded */
+               addr = (u64)pnd->logical_blk_num << DeviceInfo.nBitsInBlockDataSize;
+               ret = FTL_Replace_Block(addr);
+               if (ret == FAIL) {
+                       printk(KERN_ERR "FTL_Replace_Block fail in %s, Line %d\n", __FILE__, __LINE__);
+               }
+
+               /* Write back the updated data into NAND */
+               phy_blk = pbt[pnd->logical_blk_num] & (~BAD_BLOCK);
+               if (FAIL == GLOB_LLD_Write_Page_Main(cache_l2_blk_buf, phy_blk, 0, DeviceInfo.wPagesPerBlock)) {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "Program NAND block %d fail in %s, Line %d\n",
+                               phy_blk, __FILE__, __LINE__);
+                       /* This may not be really a bad block. So just tag it as discarded. */
+                       /* Then it has a chance to be erased when garbage collection. */
+                       /* If it is really bad, then the erase will fail and it will be marked */
+                       /* as bad then. Otherwise it will be marked as free and can be used again */
+                       MARK_BLK_AS_DISCARD(pbt[pnd->logical_blk_num]);
+                       /* Find another free block and write it again */
+                       FTL_Replace_Block(addr);
+                       phy_blk = pbt[pnd->logical_blk_num] & (~BAD_BLOCK);
+                       if (FAIL == GLOB_LLD_Write_Page_Main(cache_l2_blk_buf, phy_blk, 0, DeviceInfo.wPagesPerBlock)) {
+                               printk(KERN_ERR "Failed to write back block %d when flush L2 cache."
+                                       "Some data will be lost!\n", phy_blk);
+                               MARK_BLOCK_AS_BAD(pbt[pnd->logical_blk_num]);
+                       }
+               } else {
+                       /* tag the new free block as used block */
+                       pbt[pnd->logical_blk_num] &= (~SPARE_BLOCK);
+               }
+       }
+
+       /* Destroy the L2 Cache table and free the memory of all nodes */
+       list_for_each_entry_safe(pnd, tmp_pnd, &cache_l2.table.list, list) {
+               list_del(&pnd->list);
+               kfree(pnd);
+       }
+
+       /* Erase discard L2 cache blocks */
+       if (erase_l2_cache_blocks() != PASS)
+               nand_dbg_print(NAND_DBG_WARN,
+                       " Erase L2 cache blocks error in %s, Line %d\n",
+                       __FILE__, __LINE__);
+
+       /* Init the Level2 Cache data structure */
+       for (i = 0; i < BLK_NUM_FOR_L2_CACHE; i++)
+               cache_l2.blk_array[i] = MAX_U32_VALUE;
+       cache_l2.cur_blk_idx = 0;
+       cache_l2.cur_page_num = 0;
+       INIT_LIST_HEAD(&cache_l2.table.list);
+       cache_l2.table.logical_blk_num = MAX_U32_VALUE;
+
+       return ret;
+}
+
+/*
+ * Write back a changed victim cache item to the Level2 Cache
+ * and update the L2 Cache table to map the change.
+ * If the L2 Cache is full, then start to do the L2 Cache flush.
+*/
+static int write_back_to_l2_cache(u8 *buf, u64 logical_addr)
+{
+       u32 logical_blk_num;
+       u16 logical_page_num;
+       struct list_head *p;
+       struct spectra_l2_cache_list *pnd, *pnd_new;
+       u32 node_size;
+       int i, found;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       /*
+        * If Level2 Cache table is empty, then it means either:
+        * 1. This is the first time that the function called after FTL_init
+        * or
+        * 2. The Level2 Cache has just been flushed
+        *
+        * So, 'steal' some free blocks from NAND for L2 Cache using
+        * by just mask them as discard in the block table
+       */
+       if (list_empty(&cache_l2.table.list)) {
+               BUG_ON(cache_l2.cur_blk_idx != 0);
+               BUG_ON(cache_l2.cur_page_num!= 0);
+               BUG_ON(cache_l2.table.logical_blk_num != MAX_U32_VALUE);
+               if (FAIL == get_l2_cache_blks()) {
+                       GLOB_FTL_Garbage_Collection();
+                       if (FAIL == get_l2_cache_blks()) {
+                               printk(KERN_ALERT "Fail to get L2 cache blks!\n");
+                               return FAIL;
+                       }
+               }
+       }
+
+       logical_blk_num = BLK_FROM_ADDR(logical_addr);
+       logical_page_num = PAGE_FROM_ADDR(logical_addr, logical_blk_num);
+       BUG_ON(logical_blk_num == MAX_U32_VALUE);
+
+       /* Write the cache item data into the current position of L2 Cache */
+#if CMD_DMA
+       /*
+        * TODO
+        */
+#else
+       if (FAIL == GLOB_LLD_Write_Page_Main(buf,
+               cache_l2.blk_array[cache_l2.cur_blk_idx],
+               cache_l2.cur_page_num, 1)) {
+               nand_dbg_print(NAND_DBG_WARN, "NAND Program fail in "
+                       "%s, Line %d, new Bad Block %d generated!\n",
+                       __FILE__, __LINE__,
+                       cache_l2.blk_array[cache_l2.cur_blk_idx]);
+
+               /* TODO: tag the current block as bad and try again */
+
+               return FAIL;
+       }
+#endif
+
+       /* 
+        * Update the L2 Cache table.
+        *
+        * First seaching in the table to see whether the logical block
+        * has been mapped. If not, then kmalloc a new node for the
+        * logical block, fill data, and then insert it to the list.
+        * Otherwise, just update the mapped node directly.
+        */
+       found = 0;
+       list_for_each(p, &cache_l2.table.list) {
+               pnd = list_entry(p, struct spectra_l2_cache_list, list);
+               if (pnd->logical_blk_num == logical_blk_num) {
+                       pnd->pages_array[logical_page_num] =
+                               (cache_l2.cur_blk_idx << 16) |
+                               cache_l2.cur_page_num;
+                       found = 1;
+                       break;
+               }
+       }
+       if (!found) { /* Create new node for the logical block here */
+
+               /* The logical pages to physical pages map array is
+                * located at the end of struct spectra_l2_cache_list.
+                */ 
+               node_size = sizeof(struct spectra_l2_cache_list) +
+                       sizeof(u32) * DeviceInfo.wPagesPerBlock;
+               pnd_new = kmalloc(node_size, GFP_ATOMIC);
+               if (!pnd_new) {
+                       printk(KERN_ERR "Failed to kmalloc in %s Line %d\n",
+                               __FILE__, __LINE__);
+                       /* 
+                        * TODO: Need to flush all the L2 cache into NAND ASAP
+                        * since no memory available here
+                        */
+               }
+               pnd_new->logical_blk_num = logical_blk_num;
+               for (i = 0; i < DeviceInfo.wPagesPerBlock; i++)
+                       pnd_new->pages_array[i] = MAX_U32_VALUE;
+               pnd_new->pages_array[logical_page_num] =
+                       (cache_l2.cur_blk_idx << 16) | cache_l2.cur_page_num;
+               list_add(&pnd_new->list, &cache_l2.table.list);
+       }
+
+       /* Increasing the current position pointer of the L2 Cache */
+       cache_l2.cur_page_num++;
+       if (cache_l2.cur_page_num >= DeviceInfo.wPagesPerBlock) {
+               cache_l2.cur_blk_idx++;
+               if (cache_l2.cur_blk_idx >= BLK_NUM_FOR_L2_CACHE) {
+                       /* The L2 Cache is full. Need to flush it now */
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "L2 Cache is full, will start to flush it\n");
+                       flush_l2_cache();
+               } else {
+                       cache_l2.cur_page_num = 0;
+               }
+       }
+
+       return PASS;
+}
+
+/*
+ * Seach in the Level2 Cache table to find the cache item.
+ * If find, read the data from the NAND page of L2 Cache,
+ * Otherwise, return FAIL.
+ */
+static int search_l2_cache(u8 *buf, u64 logical_addr)
+{
+       u32 logical_blk_num;
+       u16 logical_page_num;
+       struct list_head *p;
+       struct spectra_l2_cache_list *pnd;
+       u32 tmp = MAX_U32_VALUE;
+       u32 phy_blk;
+       u16 phy_page;
+       int ret = FAIL;
+
+       logical_blk_num = BLK_FROM_ADDR(logical_addr);
+       logical_page_num = PAGE_FROM_ADDR(logical_addr, logical_blk_num);
+
+       list_for_each(p, &cache_l2.table.list) {
+               pnd = list_entry(p, struct spectra_l2_cache_list, list);
+               if (pnd->logical_blk_num == logical_blk_num) {
+                       tmp = pnd->pages_array[logical_page_num];
+                       break;
+               }
+       }
+
+       if (tmp != MAX_U32_VALUE) { /* Found valid map */
+               phy_blk = cache_l2.blk_array[(tmp >> 16) & 0xFFFF];
+               phy_page = tmp & 0xFFFF;
+#if CMD_DMA
+               /* TODO */
+#else
+               ret = GLOB_LLD_Read_Page_Main(buf, phy_blk, phy_page, 1);
+#endif
+       }
+
+       return ret;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Write_Back
+* Inputs:       pointer to data cached in sys memory
+*               address of free block in flash
+* Outputs:      PASS=0 / FAIL=1
+* Description:  writes all the pages of Cache Block to flash
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Write_Back(u8 *pData, u64 blk_addr)
+{
+       int i, j, iErase;
+       u64 old_page_addr, addr, phy_addr;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 lba;
+       
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       old_page_addr = FTL_Get_Physical_Block_Addr(blk_addr) +
+               GLOB_u64_Remainder(blk_addr, 2);
+
+       iErase = (FAIL == FTL_Replace_Block(blk_addr)) ? PASS : FAIL;
+
+       pbt[BLK_FROM_ADDR(blk_addr)] &= (~SPARE_BLOCK);
+
+#if CMD_DMA
+       p_BTableChangesDelta = (struct BTableChangesDelta *)g_pBTDelta_Free;
+       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+       p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+       p_BTableChangesDelta->BT_Index = (u32)(blk_addr >>
+               DeviceInfo.nBitsInBlockDataSize);
+       p_BTableChangesDelta->BT_Entry_Value =
+               pbt[(u32)(blk_addr >> DeviceInfo.nBitsInBlockDataSize)];
+       p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+
+       if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+               g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+               FTL_Write_IN_Progress_Block_Table_Page();
+       }
+
+       for (i = 0; i < RETRY_TIMES; i++) {
+               if (PASS == iErase) {
+                       phy_addr = FTL_Get_Physical_Block_Addr(blk_addr);
+                       if (FAIL == GLOB_FTL_Block_Erase(phy_addr)) {
+                               lba = BLK_FROM_ADDR(blk_addr);
+                               MARK_BLOCK_AS_BAD(pbt[lba]);
+                               i = RETRY_TIMES;
+                               break;
+                       }
+               }
+
+               for (j = 0; j < CACHE_ITEM_NUM; j++) {
+                       addr = Cache.array[j].address;
+                       if ((addr <= blk_addr) &&
+                               ((addr + Cache.cache_item_size) > blk_addr))
+                               cache_block_to_write = j;
+               }
+
+               phy_addr = FTL_Get_Physical_Block_Addr(blk_addr);
+               if (PASS == FTL_Cache_Update_Block(pData,
+                                       old_page_addr, phy_addr)) {
+                       cache_block_to_write = UNHIT_CACHE_ITEM;
+                       break;
+               } else {
+                       iErase = PASS;
+               }
+       }
+
+       if (i >= RETRY_TIMES) {
+               if (ERR == FTL_Flash_Error_Handle(pData,
+                                       old_page_addr, blk_addr))
+                       return ERR;
+               else
+                       return FAIL;
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Write_Page
+* Inputs:       Pointer to buffer, page address, cache block number
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It writes the data in Cache Block
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static void FTL_Cache_Write_Page(u8 *pData, u64 page_addr,
+                               u8 cache_blk, u16 flag)
+{
+       u8 *pDest;
+       u64 addr;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       addr = Cache.array[cache_blk].address;
+       pDest = Cache.array[cache_blk].buf;
+
+       pDest += (unsigned long)(page_addr - addr);
+       Cache.array[cache_blk].changed = SET;
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       int_cache[ftl_cmd_cnt].item = cache_blk;
+       int_cache[ftl_cmd_cnt].cache.address =
+                       Cache.array[cache_blk].address;
+       int_cache[ftl_cmd_cnt].cache.changed =
+                       Cache.array[cache_blk].changed;
+#endif
+       GLOB_LLD_MemCopy_CMD(pDest, pData, DeviceInfo.wPageDataSize, flag);
+       ftl_cmd_cnt++;
+#else
+       memcpy(pDest, pData, DeviceInfo.wPageDataSize);
+#endif
+       if (Cache.array[cache_blk].use_cnt < MAX_WORD_VALUE)
+               Cache.array[cache_blk].use_cnt++;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Write
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It writes least frequently used Cache block to flash if it
+*               has been changed
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Write(void)
+{
+       int i, bResult = PASS;
+       u16 bNO, least_count = 0xFFFF;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       FTL_Calculate_LRU();
+
+       bNO = Cache.LRU;
+       nand_dbg_print(NAND_DBG_DEBUG, "FTL_Cache_Write: "
+               "Least used cache block is %d\n", bNO);
+
+       if (Cache.array[bNO].changed != SET)
+               return bResult;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "FTL_Cache_Write: Cache"
+               " Block %d containing logical block %d is dirty\n",
+               bNO,
+               (u32)(Cache.array[bNO].address >>
+               DeviceInfo.nBitsInBlockDataSize));
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       int_cache[ftl_cmd_cnt].item = bNO;
+       int_cache[ftl_cmd_cnt].cache.address =
+                               Cache.array[bNO].address;
+       int_cache[ftl_cmd_cnt].cache.changed = CLEAR;
+#endif
+#endif
+       bResult = write_back_to_l2_cache(Cache.array[bNO].buf,
+                       Cache.array[bNO].address);
+       if (bResult != ERR)
+               Cache.array[bNO].changed = CLEAR;
+
+       least_count = Cache.array[bNO].use_cnt;
+
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               if (i == bNO)
+                       continue;
+               if (Cache.array[i].use_cnt > 0)
+                       Cache.array[i].use_cnt -= least_count;
+       }
+
+       return bResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Cache_Read
+* Inputs:       Page address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It reads the block from device in Cache Block
+*               Set the LRU count to 1
+*               Mark the Cache Block as clean
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Cache_Read(u64 logical_addr)
+{
+       u64 item_addr, phy_addr;
+       u16 num;
+       int ret;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       num = Cache.LRU; /* The LRU cache item will be overwritten */
+
+       item_addr = (u64)GLOB_u64_Div(logical_addr, Cache.cache_item_size) *
+               Cache.cache_item_size;
+       Cache.array[num].address = item_addr;
+       Cache.array[num].use_cnt = 1;
+       Cache.array[num].changed = CLEAR;
+
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+       int_cache[ftl_cmd_cnt].item = num;
+       int_cache[ftl_cmd_cnt].cache.address =
+                       Cache.array[num].address;
+       int_cache[ftl_cmd_cnt].cache.changed =
+                       Cache.array[num].changed;
+#endif
+#endif
+       /*
+        * Search in L2 Cache. If hit, fill data into L1 Cache item buffer,
+        * Otherwise, read it from NAND
+        */
+       ret = search_l2_cache(Cache.array[num].buf, logical_addr);
+       if (PASS == ret) /* Hit in L2 Cache */
+               return ret;
+
+       /* Compute the physical start address of NAND device according to */
+       /* the logical start address of the cache item (LRU cache item) */
+       phy_addr = FTL_Get_Physical_Block_Addr(item_addr) +
+               GLOB_u64_Remainder(item_addr, 2);
+
+       return FTL_Cache_Read_All(Cache.array[num].buf, phy_addr);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Check_Block_Table
+* Inputs:       ?
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It checks the correctness of each block table entry
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Check_Block_Table(int wOldTable)
+{
+       u32 i;
+       int wResult = PASS;
+       u32 blk_idx;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u8 *pFlag = flag_check_blk_table;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (NULL != pFlag) {
+               memset(pFlag, FAIL, DeviceInfo.wDataBlockNum);
+               for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+                       blk_idx = (u32)(pbt[i] & (~BAD_BLOCK));
+
+                       /*
+                        * 20081006/KBV - Changed to pFlag[i] reference
+                        * to avoid buffer overflow
+                        */
+
+                       /*
+                        * 2008-10-20 Yunpeng Note: This change avoid
+                        * buffer overflow, but changed function of
+                        * the code, so it should be re-write later
+                        */
+                       if ((blk_idx > DeviceInfo.wSpectraEndBlock) ||
+                               PASS == pFlag[i]) {
+                               wResult = FAIL;
+                               break;
+                       } else {
+                               pFlag[i] = PASS;
+                       }
+               }
+       }
+
+       return wResult;
+}
+
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Write_Block_Table
+* Inputs:       flasg
+* Outputs:      0=Block Table was updated. No write done. 1=Block write needs to
+* happen. -1 Error
+* Description:  It writes the block table
+*               Block table always mapped to LBA 0 which inturn mapped
+*               to any physical block
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Write_Block_Table(int wForce)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       int wSuccess = PASS;
+       u32 wTempBlockTableIndex;
+       u16 bt_pages, new_bt_offset;
+       u8 blockchangeoccured = 0;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       bt_pages = FTL_Get_Block_Table_Flash_Size_Pages();
+
+       if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus)
+               return 0;
+
+       if (PASS == wForce) {
+               g_wBlockTableOffset =
+                       (u16)(DeviceInfo.wPagesPerBlock - bt_pages);
+#if CMD_DMA
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+               p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+               p_BTableChangesDelta->g_wBlockTableOffset =
+                       g_wBlockTableOffset;
+               p_BTableChangesDelta->ValidFields = 0x01;
+#endif
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "Inside FTL_Write_Block_Table: block %d Page:%d\n",
+               g_wBlockTableIndex, g_wBlockTableOffset);
+
+       do {
+               new_bt_offset = g_wBlockTableOffset + bt_pages + 1;
+               if ((0 == (new_bt_offset % DeviceInfo.wPagesPerBlock)) ||
+                       (new_bt_offset > DeviceInfo.wPagesPerBlock) ||
+                       (FAIL == wSuccess)) {
+                       wTempBlockTableIndex = FTL_Replace_Block_Table();
+                       if (BAD_BLOCK == wTempBlockTableIndex)
+                               return ERR;
+                       if (!blockchangeoccured) {
+                               bt_block_changed = 1;
+                               blockchangeoccured = 1;
+                       }
+
+                       g_wBlockTableIndex = wTempBlockTableIndex;
+                       g_wBlockTableOffset = 0;
+                       pbt[BLOCK_TABLE_INDEX] = g_wBlockTableIndex;
+#if CMD_DMA
+                       p_BTableChangesDelta =
+                               (struct BTableChangesDelta *)g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                   ftl_cmd_cnt;
+                       p_BTableChangesDelta->g_wBlockTableOffset =
+                                   g_wBlockTableOffset;
+                       p_BTableChangesDelta->g_wBlockTableIndex =
+                                   g_wBlockTableIndex;
+                       p_BTableChangesDelta->ValidFields = 0x03;
+
+                       p_BTableChangesDelta =
+                               (struct BTableChangesDelta *)g_pBTDelta_Free;
+                       g_pBTDelta_Free +=
+                               sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                   ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index =
+                                   BLOCK_TABLE_INDEX;
+                       p_BTableChangesDelta->BT_Entry_Value =
+                                   pbt[BLOCK_TABLE_INDEX];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+               }
+
+               wSuccess = FTL_Write_Block_Table_Data();
+               if (FAIL == wSuccess)
+                       MARK_BLOCK_AS_BAD(pbt[BLOCK_TABLE_INDEX]);
+       } while (FAIL == wSuccess);
+
+       g_cBlockTableStatus = CURRENT_BLOCK_TABLE;
+
+       return 1;
+}
+
+/******************************************************************
+* Function:     GLOB_FTL_Flash_Format
+* Inputs:       none
+* Outputs:      PASS
+* Description:  The block table stores bad block info, including MDF+
+*               blocks gone bad over the ages. Therefore, if we have a
+*               block table in place, then use it to scan for bad blocks
+*               If not, then scan for MDF.
+*               Now, a block table will only be found if spectra was already
+*               being used. For a fresh flash, we'll go thru scanning for
+*               MDF. If spectra was being used, then there is a chance that
+*               the MDF has been corrupted. Spectra avoids writing to the
+*               first 2 bytes of the spare area to all pages in a block. This
+*               covers all known flash devices. However, since flash
+*               manufacturers have no standard of where the MDF is stored,
+*               this cannot guarantee that the MDF is protected for future
+*               devices too. The initial scanning for the block table assures
+*               this. It is ok even if the block table is outdated, as all
+*               we're looking for are bad block markers.
+*               Use this when mounting a file system or starting a
+*               new flash.
+*
+*********************************************************************/
+static int  FTL_Format_Flash(u8 valid_block_table)
+{
+       u32 i, j;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 tempNode;
+       int ret;
+
+#if CMD_DMA
+       u32 *pbtStartingCopy = (u32 *)g_pBTStartingCopy;
+       if (ftl_cmd_cnt)
+               return FAIL;
+#endif
+
+       if (FAIL == FTL_Check_Block_Table(FAIL))
+               valid_block_table = 0;
+
+       if (valid_block_table) {
+               u8 switched = 1;
+               u32 block, k;
+
+               k = DeviceInfo.wSpectraStartBlock;
+               while (switched && (k < DeviceInfo.wSpectraEndBlock)) {
+                       switched = 0;
+                       k++;
+                       for (j = DeviceInfo.wSpectraStartBlock, i = 0;
+                       j <= DeviceInfo.wSpectraEndBlock;
+                       j++, i++) {
+                               block = (pbt[i] & ~BAD_BLOCK) -
+                                       DeviceInfo.wSpectraStartBlock;
+                               if (block != i) {
+                                       switched = 1;
+                                       tempNode = pbt[i];
+                                       pbt[i] = pbt[block];
+                                       pbt[block] = tempNode;
+                               }
+                       }
+               }
+               if ((k == DeviceInfo.wSpectraEndBlock) && switched)
+                       valid_block_table = 0;
+       }
+
+       if (!valid_block_table) {
+               memset(g_pBlockTable, 0,
+                       DeviceInfo.wDataBlockNum * sizeof(u32));
+               memset(g_pWearCounter, 0,
+                       DeviceInfo.wDataBlockNum * sizeof(u8));
+               if (DeviceInfo.MLCDevice)
+                       memset(g_pReadCounter, 0,
+                               DeviceInfo.wDataBlockNum * sizeof(u16));
+#if CMD_DMA
+               memset(g_pBTStartingCopy, 0,
+                       DeviceInfo.wDataBlockNum * sizeof(u32));
+               memset(g_pWearCounterCopy, 0,
+                               DeviceInfo.wDataBlockNum * sizeof(u8));
+               if (DeviceInfo.MLCDevice)
+                       memset(g_pReadCounterCopy, 0,
+                               DeviceInfo.wDataBlockNum * sizeof(u16));
+#endif
+               for (j = DeviceInfo.wSpectraStartBlock, i = 0;
+                       j <= DeviceInfo.wSpectraEndBlock;
+                       j++, i++) {
+                       if (GLOB_LLD_Get_Bad_Block((u32)j))
+                               pbt[i] = (u32)(BAD_BLOCK | j);
+               }
+       }
+
+       nand_dbg_print(NAND_DBG_WARN, "Erasing all blocks in the NAND\n");
+
+       for (j = DeviceInfo.wSpectraStartBlock, i = 0;
+               j <= DeviceInfo.wSpectraEndBlock;
+               j++, i++) {
+               if ((pbt[i] & BAD_BLOCK) != BAD_BLOCK) {
+                       ret = GLOB_LLD_Erase_Block(j);
+                       if (FAIL == ret) {
+                               pbt[i] = (u32)(j);
+                               MARK_BLOCK_AS_BAD(pbt[i]);
+                               nand_dbg_print(NAND_DBG_WARN,
+                              "NAND Program fail in %s, Line %d, "
+                              "Function: %s, new Bad Block %d generated!\n",
+                              __FILE__, __LINE__, __func__, (int)j);
+                       } else {
+                               pbt[i] = (u32)(SPARE_BLOCK | j);
+                       }
+               }
+#if CMD_DMA
+               pbtStartingCopy[i] = pbt[i];
+#endif
+       }
+
+       g_wBlockTableOffset = 0;
+       for (i = 0; (i <= (DeviceInfo.wSpectraEndBlock -
+                       DeviceInfo.wSpectraStartBlock))
+                       && ((pbt[i] & BAD_BLOCK) == BAD_BLOCK); i++)
+               ;
+       if (i > (DeviceInfo.wSpectraEndBlock - DeviceInfo.wSpectraStartBlock)) {
+               printk(KERN_ERR "All blocks bad!\n");
+               return FAIL;
+       } else {
+               g_wBlockTableIndex = pbt[i] & ~BAD_BLOCK;
+               if (i != BLOCK_TABLE_INDEX) {
+                       tempNode = pbt[i];
+                       pbt[i] = pbt[BLOCK_TABLE_INDEX];
+                       pbt[BLOCK_TABLE_INDEX] = tempNode;
+               }
+       }
+       pbt[BLOCK_TABLE_INDEX] &= (~SPARE_BLOCK);
+
+#if CMD_DMA
+       pbtStartingCopy[BLOCK_TABLE_INDEX] &= (~SPARE_BLOCK);
+#endif
+
+       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+       memset(g_pBTBlocks, 0xFF,
+                       (1 + LAST_BT_ID - FIRST_BT_ID) * sizeof(u32));
+       g_pBTBlocks[FIRST_BT_ID-FIRST_BT_ID] = g_wBlockTableIndex;
+       FTL_Write_Block_Table(FAIL);
+
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               Cache.array[i].address = NAND_CACHE_INIT_ADDR;
+               Cache.array[i].use_cnt = 0;
+               Cache.array[i].changed  = CLEAR;
+       }
+
+#if (RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE && CMD_DMA)
+       memcpy((void *)&cache_start_copy, (void *)&Cache,
+                       sizeof(struct flash_cache_tag));
+#endif
+       return PASS;
+}
+
+static int  force_format_nand(void)
+{
+       u32 i;
+
+       /* Force erase the whole unprotected physical partiton of NAND */
+       printk(KERN_ALERT "Start to force erase whole NAND device ...\n");
+       printk(KERN_ALERT "From phyical block %d to %d\n",
+               DeviceInfo.wSpectraStartBlock, DeviceInfo.wSpectraEndBlock);
+       for (i = DeviceInfo.wSpectraStartBlock; i <= DeviceInfo.wSpectraEndBlock; i++) {
+               if (GLOB_LLD_Erase_Block(i))
+                       printk(KERN_ERR "Failed to force erase NAND block %d\n", i);
+       }
+       printk(KERN_ALERT "Force Erase ends. Please reboot the system ...\n");
+       while(1);
+
+       return PASS;
+}
+
+int GLOB_FTL_Flash_Format(void)
+{
+       //return FTL_Format_Flash(1);
+       return force_format_nand();
+
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Search_Block_Table_IN_Block
+* Inputs:       Block Number
+*               Pointer to page
+* Outputs:      PASS / FAIL
+*               Page contatining the block table
+* Description:  It searches the block table in the block
+*               passed as an argument.
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Search_Block_Table_IN_Block(u32 BT_Block,
+                                               u8 BT_Tag, u16 *Page)
+{
+       u16 i, j, k;
+       u16 Result = PASS;
+       u16 Last_IPF = 0;
+       u8  BT_Found = 0;
+       u8 *tagarray;
+       u8 *tempbuf = tmp_buf_search_bt_in_block;
+       u8 *pSpareBuf = spare_buf_search_bt_in_block;
+       u8 *pSpareBufBTLastPage = spare_buf_bt_search_bt_in_block;
+       u8 bt_flag_last_page = 0xFF;
+       u8 search_in_previous_pages = 0;
+       u16 bt_pages;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+                      "Searching block table in %u block\n",
+                      (unsigned int)BT_Block);
+
+       bt_pages = FTL_Get_Block_Table_Flash_Size_Pages();
+
+       for (i = bt_pages; i < DeviceInfo.wPagesPerBlock;
+                               i += (bt_pages + 1)) {
+               nand_dbg_print(NAND_DBG_DEBUG,
+                              "Searching last IPF: %d\n", i);
+               Result = GLOB_LLD_Read_Page_Main_Polling(tempbuf,
+                                                       BT_Block, i, 1);
+
+               if (0 == memcmp(tempbuf, g_pIPF, DeviceInfo.wPageDataSize)) {
+                       if ((i + bt_pages + 1) < DeviceInfo.wPagesPerBlock) {
+                               continue;
+                       } else {
+                               search_in_previous_pages = 1;
+                               Last_IPF = i;
+                       }
+               }
+
+               if (!search_in_previous_pages) {
+                       if (i != bt_pages) {
+                               i -= (bt_pages + 1);
+                               Last_IPF = i;
+                       }
+               }
+
+               if (0 == Last_IPF)
+                       break;
+
+               if (!search_in_previous_pages) {
+                       i = i + 1;
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Reading the spare area of Block %u Page %u",
+                               (unsigned int)BT_Block, i);
+                       Result = GLOB_LLD_Read_Page_Spare(pSpareBuf,
+                                                       BT_Block, i, 1);
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Reading the spare area of Block %u Page %u",
+                               (unsigned int)BT_Block, i + bt_pages - 1);
+                       Result = GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage,
+                               BT_Block, i + bt_pages - 1, 1);
+
+                       k = 0;
+                       j = FTL_Extract_Block_Table_Tag(pSpareBuf, &tagarray);
+                       if (j) {
+                               for (; k < j; k++) {
+                                       if (tagarray[k] == BT_Tag)
+                                               break;
+                               }
+                       }
+
+                       if (k < j)
+                               bt_flag = tagarray[k];
+                       else
+                               Result = FAIL;
+
+                       if (Result == PASS) {
+                               k = 0;
+                               j = FTL_Extract_Block_Table_Tag(
+                                       pSpareBufBTLastPage, &tagarray);
+                               if (j) {
+                                       for (; k < j; k++) {
+                                               if (tagarray[k] == BT_Tag)
+                                                       break;
+                                       }
+                               }
+
+                               if (k < j)
+                                       bt_flag_last_page = tagarray[k];
+                               else
+                                       Result = FAIL;
+
+                               if (Result == PASS) {
+                                       if (bt_flag == bt_flag_last_page) {
+                                               nand_dbg_print(NAND_DBG_DEBUG,
+                                                       "Block table is found"
+                                                       " in page after IPF "
+                                                       "at block %d "
+                                                       "page %d\n",
+                                                       (int)BT_Block, i);
+                                               BT_Found = 1;
+                                               *Page  = i;
+                                               g_cBlockTableStatus =
+                                                       CURRENT_BLOCK_TABLE;
+                                               break;
+                                       } else {
+                                               Result = FAIL;
+                                       }
+                               }
+                       }
+               }
+
+               if (search_in_previous_pages)
+                       i = i - bt_pages;
+               else
+                       i = i - (bt_pages + 1);
+
+               Result = PASS;
+
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Reading the spare area of Block %d Page %d",
+                       (int)BT_Block, i);
+
+               Result = GLOB_LLD_Read_Page_Spare(pSpareBuf, BT_Block, i, 1);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Reading the spare area of Block %u Page %u",
+                       (unsigned int)BT_Block, i + bt_pages - 1);
+
+               Result = GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage,
+                                       BT_Block, i + bt_pages - 1, 1);
+
+               k = 0;
+               j = FTL_Extract_Block_Table_Tag(pSpareBuf, &tagarray);
+               if (j) {
+                       for (; k < j; k++) {
+                               if (tagarray[k] == BT_Tag)
+                                       break;
+                       }
+               }
+
+               if (k < j)
+                       bt_flag = tagarray[k];
+               else
+                       Result = FAIL;
+
+               if (Result == PASS) {
+                       k = 0;
+                       j = FTL_Extract_Block_Table_Tag(pSpareBufBTLastPage,
+                                               &tagarray);
+                       if (j) {
+                               for (; k < j; k++) {
+                                       if (tagarray[k] == BT_Tag)
+                                               break;
+                               }
+                       }
+
+                       if (k < j) {
+                               bt_flag_last_page = tagarray[k];
+                       } else {
+                               Result = FAIL;
+                               break;
+                       }
+
+                       if (Result == PASS) {
+                               if (bt_flag == bt_flag_last_page) {
+                                       nand_dbg_print(NAND_DBG_DEBUG,
+                                               "Block table is found "
+                                               "in page prior to IPF "
+                                               "at block %u page %d\n",
+                                               (unsigned int)BT_Block, i);
+                                       BT_Found = 1;
+                                       *Page  = i;
+                                       g_cBlockTableStatus =
+                                               IN_PROGRESS_BLOCK_TABLE;
+                                       break;
+                               } else {
+                                       Result = FAIL;
+                                       break;
+                               }
+                       }
+               }
+       }
+
+       if (Result == FAIL) {
+               if ((Last_IPF > bt_pages) && (i < Last_IPF) && (!BT_Found)) {
+                       BT_Found = 1;
+                       *Page = i - (bt_pages + 1);
+               }
+               if ((Last_IPF == bt_pages) && (i < Last_IPF) && (!BT_Found))
+                       goto func_return;
+       }
+
+       if (Last_IPF == 0) {
+               i = 0;
+               Result = PASS;
+               nand_dbg_print(NAND_DBG_DEBUG, "Reading the spare area of "
+                       "Block %u Page %u", (unsigned int)BT_Block, i);
+
+               Result = GLOB_LLD_Read_Page_Spare(pSpareBuf, BT_Block, i, 1);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Reading the spare area of Block %u Page %u",
+                       (unsigned int)BT_Block, i + bt_pages - 1);
+               Result = GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage,
+                                       BT_Block, i + bt_pages - 1, 1);
+
+               k = 0;
+               j = FTL_Extract_Block_Table_Tag(pSpareBuf, &tagarray);
+               if (j) {
+                       for (; k < j; k++) {
+                               if (tagarray[k] == BT_Tag)
+                                       break;
+                       }
+               }
+
+               if (k < j)
+                       bt_flag = tagarray[k];
+               else
+                       Result = FAIL;
+
+               if (Result == PASS) {
+                       k = 0;
+                       j = FTL_Extract_Block_Table_Tag(pSpareBufBTLastPage,
+                                                       &tagarray);
+                       if (j) {
+                               for (; k < j; k++) {
+                                       if (tagarray[k] == BT_Tag)
+                                               break;
+                               }
+                       }
+
+                       if (k < j)
+                               bt_flag_last_page = tagarray[k];
+                       else
+                               Result = FAIL;
+
+                       if (Result == PASS) {
+                               if (bt_flag == bt_flag_last_page) {
+                                       nand_dbg_print(NAND_DBG_DEBUG,
+                                               "Block table is found "
+                                               "in page after IPF at "
+                                               "block %u page %u\n",
+                                               (unsigned int)BT_Block,
+                                               (unsigned int)i);
+                                       BT_Found = 1;
+                                       *Page  = i;
+                                       g_cBlockTableStatus =
+                                               CURRENT_BLOCK_TABLE;
+                                       goto func_return;
+                               } else {
+                                       Result = FAIL;
+                               }
+                       }
+               }
+
+               if (Result == FAIL)
+                       goto func_return;
+       }
+func_return:
+       return Result;
+}
+
+u8 *get_blk_table_start_addr(void)
+{
+       return g_pBlockTable;
+}
+
+unsigned long get_blk_table_len(void)
+{
+       return DeviceInfo.wDataBlockNum * sizeof(u32);
+}
+
+u8 *get_wear_leveling_table_start_addr(void)
+{
+       return g_pWearCounter;
+}
+
+unsigned long get_wear_leveling_table_len(void)
+{
+       return DeviceInfo.wDataBlockNum * sizeof(u8);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Read_Block_Table
+* Inputs:       none
+* Outputs:      PASS / FAIL
+* Description:  read the flash spare area and find a block containing the
+*               most recent block table(having largest block_table_counter).
+*               Find the last written Block table in this block.
+*               Check the correctness of Block Table
+*               If CDMA is enabled, this function is called in
+*               polling mode.
+*               We don't need to store changes in Block table in this
+*               function as it is called only at initialization
+*
+*               Note: Currently this function is called at initialization
+*               before any read/erase/write command issued to flash so,
+*               there is no need to wait for CDMA list to complete as of now
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Read_Block_Table(void)
+{
+       u16 i = 0;
+       int k, j;
+       u8 *tempBuf, *tagarray;
+       int wResult = FAIL;
+       int status = FAIL;
+       u8 block_table_found = 0;
+       int search_result;
+       u32 Block;
+       u16 Page = 0;
+       u16 PageCount;
+       u16 bt_pages;
+       int wBytesCopied = 0, tempvar;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       tempBuf = tmp_buf1_read_blk_table;
+       bt_pages = FTL_Get_Block_Table_Flash_Size_Pages();
+
+       for (j = DeviceInfo.wSpectraStartBlock;
+               j <= (int)DeviceInfo.wSpectraEndBlock;
+                       j++) {
+               status = GLOB_LLD_Read_Page_Spare(tempBuf, j, 0, 1);
+               k = 0;
+               i = FTL_Extract_Block_Table_Tag(tempBuf, &tagarray);
+               if (i) {
+                       status  = GLOB_LLD_Read_Page_Main_Polling(tempBuf,
+                                                               j, 0, 1);
+                       for (; k < i; k++) {
+                               if (tagarray[k] == tempBuf[3])
+                                       break;
+                       }
+               }
+
+               if (k < i)
+                       k = tagarray[k];
+               else
+                       continue;
+
+               nand_dbg_print(NAND_DBG_DEBUG,
+                               "Block table is contained in Block %d %d\n",
+                                      (unsigned int)j, (unsigned int)k);
+
+               if (g_pBTBlocks[k-FIRST_BT_ID] == BTBLOCK_INVAL) {
+                       g_pBTBlocks[k-FIRST_BT_ID] = j;
+                       block_table_found = 1;
+               } else {
+                       printk(KERN_ERR "FTL_Read_Block_Table -"
+                               "This should never happens. "
+                               "Two block table have same counter %u!\n", k);
+               }
+       }
+
+       if (block_table_found) {
+               if (g_pBTBlocks[FIRST_BT_ID - FIRST_BT_ID] != BTBLOCK_INVAL &&
+               g_pBTBlocks[LAST_BT_ID - FIRST_BT_ID] != BTBLOCK_INVAL) {
+                       j = LAST_BT_ID;
+                       while ((j > FIRST_BT_ID) &&
+                       (g_pBTBlocks[j - FIRST_BT_ID] != BTBLOCK_INVAL))
+                               j--;
+                       if (j == FIRST_BT_ID) {
+                               j = LAST_BT_ID;
+                               last_erased = LAST_BT_ID;
+                       } else {
+                               last_erased = (u8)j + 1;
+                               while ((j > FIRST_BT_ID) && (BTBLOCK_INVAL ==
+                                       g_pBTBlocks[j - FIRST_BT_ID]))
+                                       j--;
+                       }
+               } else {
+                       j = FIRST_BT_ID;
+                       while (g_pBTBlocks[j - FIRST_BT_ID] == BTBLOCK_INVAL)
+                               j++;
+                       last_erased = (u8)j;
+                       while ((j < LAST_BT_ID) && (BTBLOCK_INVAL !=
+                               g_pBTBlocks[j - FIRST_BT_ID]))
+                               j++;
+                       if (g_pBTBlocks[j-FIRST_BT_ID] == BTBLOCK_INVAL)
+                               j--;
+               }
+
+               if (last_erased > j)
+                       j += (1 + LAST_BT_ID - FIRST_BT_ID);
+
+               for (; (j >= last_erased) && (FAIL == wResult); j--) {
+                       i = (j - FIRST_BT_ID) %
+                               (1 + LAST_BT_ID - FIRST_BT_ID);
+                       search_result =
+                       FTL_Search_Block_Table_IN_Block(g_pBTBlocks[i],
+                                               i + FIRST_BT_ID, &Page);
+                       if (g_cBlockTableStatus == IN_PROGRESS_BLOCK_TABLE)
+                               block_table_found = 0;
+
+                       while ((search_result == PASS) && (FAIL == wResult)) {
+                               nand_dbg_print(NAND_DBG_DEBUG,
+                                       "FTL_Read_Block_Table:"
+                                       "Block: %u Page: %u "
+                                       "contains block table\n",
+                                       (unsigned int)g_pBTBlocks[i],
+                                       (unsigned int)Page);
+
+                               tempBuf = tmp_buf2_read_blk_table;
+
+                               for (k = 0; k < bt_pages; k++) {
+                                       Block = g_pBTBlocks[i];
+                                       PageCount = 1;
+
+                                       status  =
+                                       GLOB_LLD_Read_Page_Main_Polling(
+                                       tempBuf, Block, Page, PageCount);
+
+                                       tempvar = k ? 0 : 4;
+
+                                       wBytesCopied +=
+                                       FTL_Copy_Block_Table_From_Flash(
+                                       tempBuf + tempvar,
+                                       DeviceInfo.wPageDataSize - tempvar,
+                                       wBytesCopied);
+
+                                       Page++;
+                               }
+
+                               wResult = FTL_Check_Block_Table(FAIL);
+                               if (FAIL == wResult) {
+                                       block_table_found = 0;
+                                       if (Page > bt_pages)
+                                               Page -= ((bt_pages<<1) + 1);
+                                       else
+                                               search_result = FAIL;
+                               }
+                       }
+               }
+       }
+
+       if (PASS == wResult) {
+               if (!block_table_found)
+                       FTL_Execute_SPL_Recovery();
+
+               if (g_cBlockTableStatus == IN_PROGRESS_BLOCK_TABLE)
+                       g_wBlockTableOffset = (u16)Page + 1;
+               else
+                       g_wBlockTableOffset = (u16)Page - bt_pages;
+
+               g_wBlockTableIndex = (u32)g_pBTBlocks[i];
+
+#if CMD_DMA
+               if (DeviceInfo.MLCDevice)
+                       memcpy(g_pBTStartingCopy, g_pBlockTable,
+                               DeviceInfo.wDataBlockNum * sizeof(u32)
+                               + DeviceInfo.wDataBlockNum * sizeof(u8)
+                               + DeviceInfo.wDataBlockNum * sizeof(u16));
+               else
+                       memcpy(g_pBTStartingCopy, g_pBlockTable,
+                               DeviceInfo.wDataBlockNum * sizeof(u32)
+                               + DeviceInfo.wDataBlockNum * sizeof(u8));
+#endif
+       }
+
+       if (FAIL == wResult)
+               printk(KERN_ERR "Yunpeng - "
+               "Can not find valid spectra block table!\n");
+
+#if AUTO_FORMAT_FLASH
+       if (FAIL == wResult) {
+               nand_dbg_print(NAND_DBG_DEBUG, "doing auto-format\n");
+               wResult = FTL_Format_Flash(0);
+       }
+#endif
+
+       return wResult;
+}
+
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Flash_Error_Handle
+* Inputs:       Pointer to data
+*               Page address
+*               Block address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  It handles any error occured during Spectra operation
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Flash_Error_Handle(u8 *pData, u64 old_page_addr,
+                               u64 blk_addr)
+{
+       u32 i;
+       int j;
+       u32 tmp_node, blk_node = BLK_FROM_ADDR(blk_addr);
+       u64 phy_addr;
+       int wErase = FAIL;
+       int wResult = FAIL;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (ERR == GLOB_FTL_Garbage_Collection())
+               return ERR;
+
+       do {
+               for (i = DeviceInfo.wSpectraEndBlock -
+                       DeviceInfo.wSpectraStartBlock;
+                                       i > 0; i--) {
+                       if (IS_SPARE_BLOCK(i)) {
+                               tmp_node = (u32)(BAD_BLOCK |
+                                       pbt[blk_node]);
+                               pbt[blk_node] = (u32)(pbt[i] &
+                                       (~SPARE_BLOCK));
+                               pbt[i] = tmp_node;
+#if CMD_DMA
+                               p_BTableChangesDelta =
+                                   (struct BTableChangesDelta *)
+                                   g_pBTDelta_Free;
+                               g_pBTDelta_Free +=
+                                   sizeof(struct BTableChangesDelta);
+
+                               p_BTableChangesDelta->ftl_cmd_cnt =
+                                   ftl_cmd_cnt;
+                               p_BTableChangesDelta->BT_Index =
+                                   blk_node;
+                               p_BTableChangesDelta->BT_Entry_Value =
+                                   pbt[blk_node];
+                               p_BTableChangesDelta->ValidFields = 0x0C;
+
+                               p_BTableChangesDelta =
+                                   (struct BTableChangesDelta *)
+                                   g_pBTDelta_Free;
+                               g_pBTDelta_Free +=
+                                   sizeof(struct BTableChangesDelta);
+
+                               p_BTableChangesDelta->ftl_cmd_cnt =
+                                   ftl_cmd_cnt;
+                               p_BTableChangesDelta->BT_Index = i;
+                               p_BTableChangesDelta->BT_Entry_Value = pbt[i];
+                               p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+                               wResult = PASS;
+                               break;
+                       }
+               }
+
+               if (FAIL == wResult) {
+                       if (FAIL == GLOB_FTL_Garbage_Collection())
+                               break;
+                       else
+                               continue;
+               }
+
+               if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+
+               phy_addr = FTL_Get_Physical_Block_Addr(blk_addr);
+
+               for (j = 0; j < RETRY_TIMES; j++) {
+                       if (PASS == wErase) {
+                               if (FAIL == GLOB_FTL_Block_Erase(phy_addr)) {
+                                       MARK_BLOCK_AS_BAD(pbt[blk_node]);
+                                       break;
+                               }
+                       }
+                       if (PASS == FTL_Cache_Update_Block(pData,
+                                                          old_page_addr,
+                                                          phy_addr)) {
+                               wResult = PASS;
+                               break;
+                       } else {
+                               wResult = FAIL;
+                               wErase = PASS;
+                       }
+               }
+       } while (FAIL == wResult);
+
+       FTL_Write_Block_Table(FAIL);
+
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Get_Page_Num
+* Inputs:       Size in bytes
+* Outputs:      Size in pages
+* Description:  It calculates the pages required for the length passed
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Get_Page_Num(u64 length)
+{
+       return (u32)((length >> DeviceInfo.nBitsInPageDataSize) +
+               (GLOB_u64_Remainder(length , 1) > 0 ? 1 : 0));
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Get_Physical_Block_Addr
+* Inputs:       Block Address (byte format)
+* Outputs:      Physical address of the block.
+* Description:  It translates LBA to PBA by returning address stored
+*               at the LBA location in the block table
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u64 FTL_Get_Physical_Block_Addr(u64 logical_addr)
+{
+       u32 *pbt;
+       u64 physical_addr;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       pbt = (u32 *)g_pBlockTable;
+       physical_addr = (u64) DeviceInfo.wBlockDataSize *
+               (pbt[BLK_FROM_ADDR(logical_addr)] & (~BAD_BLOCK));
+
+       return physical_addr;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Get_Block_Index
+* Inputs:       Physical Block no.
+* Outputs:      Logical block no. /BAD_BLOCK
+* Description:  It returns the logical block no. for the PBA passed
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Get_Block_Index(u32 wBlockNum)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++)
+               if (wBlockNum == (pbt[i] & (~BAD_BLOCK)))
+                       return i;
+
+       return BAD_BLOCK;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Wear_Leveling
+* Inputs:       none
+* Outputs:      PASS=0
+* Description:  This is static wear leveling (done by explicit call)
+*               do complete static wear leveling
+*               do complete garbage collection
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Wear_Leveling(void)
+{
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       FTL_Static_Wear_Leveling();
+       GLOB_FTL_Garbage_Collection();
+
+       return PASS;
+}
+
+static void find_least_most_worn(u8 *chg,
+       u32 *least_idx, u8 *least_cnt,
+       u32 *most_idx, u8 *most_cnt)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 idx;
+       u8 cnt;
+       int i;
+
+       for (i = BLOCK_TABLE_INDEX + 1; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_BAD_BLOCK(i) || PASS == chg[i])
+                       continue;
+
+               idx = (u32) ((~BAD_BLOCK) & pbt[i]);
+               cnt = g_pWearCounter[idx - DeviceInfo.wSpectraStartBlock];
+
+               if (IS_SPARE_BLOCK(i)) {
+                       if (cnt > *most_cnt) {
+                               *most_cnt = cnt;
+                               *most_idx = idx;
+                       }
+               }
+
+               if (IS_DATA_BLOCK(i)) {
+                       if (cnt < *least_cnt) {
+                               *least_cnt = cnt;
+                               *least_idx = idx;
+                       }
+               }
+
+               if (PASS == chg[*most_idx] || PASS == chg[*least_idx]) {
+                       debug_boundary_error(*most_idx,
+                               DeviceInfo.wDataBlockNum, 0);
+                       debug_boundary_error(*least_idx,
+                               DeviceInfo.wDataBlockNum, 0);
+                       continue;
+               }
+       }
+}
+
+static int move_blks_for_wear_leveling(u8 *chg,
+       u32 *least_idx, u32 *rep_blk_num, int *result)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 rep_blk;
+       int j, ret_cp_blk, ret_erase;
+       int ret = PASS;
+
+       chg[*least_idx] = PASS;
+       debug_boundary_error(*least_idx, DeviceInfo.wDataBlockNum, 0);
+
+       rep_blk = FTL_Replace_MWBlock();
+       if (rep_blk != BAD_BLOCK) {
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "More than two spare blocks exist so do it\n");
+               nand_dbg_print(NAND_DBG_DEBUG, "Block Replaced is %d\n",
+                               rep_blk);
+
+               chg[rep_blk] = PASS;
+
+               if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+
+               for (j = 0; j < RETRY_TIMES; j++) {
+                       ret_cp_blk = FTL_Copy_Block((u64)(*least_idx) *
+                               DeviceInfo.wBlockDataSize,
+                               (u64)rep_blk * DeviceInfo.wBlockDataSize);
+                       if (FAIL == ret_cp_blk) {
+                               ret_erase = GLOB_FTL_Block_Erase((u64)rep_blk
+                                       * DeviceInfo.wBlockDataSize);
+                               if (FAIL == ret_erase)
+                                       MARK_BLOCK_AS_BAD(pbt[rep_blk]);
+                       } else {
+                               nand_dbg_print(NAND_DBG_DEBUG,
+                                       "FTL_Copy_Block == OK\n");
+                               break;
+                       }
+               }
+
+               if (j < RETRY_TIMES) {
+                       u32 tmp;
+                       u32 old_idx = FTL_Get_Block_Index(*least_idx);
+                       u32 rep_idx = FTL_Get_Block_Index(rep_blk);
+                       tmp = (u32)(DISCARD_BLOCK | pbt[old_idx]);
+                       pbt[old_idx] = (u32)((~SPARE_BLOCK) &
+                                                       pbt[rep_idx]);
+                       pbt[rep_idx] = tmp;
+#if CMD_DMA
+                       p_BTableChangesDelta = (struct BTableChangesDelta *)
+                                               g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                               ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index = old_idx;
+                       p_BTableChangesDelta->BT_Entry_Value = pbt[old_idx];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+
+                       p_BTableChangesDelta = (struct BTableChangesDelta *)
+                                               g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                               ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index = rep_idx;
+                       p_BTableChangesDelta->BT_Entry_Value = pbt[rep_idx];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+               } else {
+                       pbt[FTL_Get_Block_Index(rep_blk)] |= BAD_BLOCK;
+#if CMD_DMA
+                       p_BTableChangesDelta = (struct BTableChangesDelta *)
+                                               g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                               ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index =
+                                       FTL_Get_Block_Index(rep_blk);
+                       p_BTableChangesDelta->BT_Entry_Value =
+                                       pbt[FTL_Get_Block_Index(rep_blk)];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+                       *result = FAIL;
+                       ret = FAIL;
+               }
+
+               if (((*rep_blk_num)++) > WEAR_LEVELING_BLOCK_NUM)
+                       ret = FAIL;
+       } else {
+               printk(KERN_ERR "Less than 3 spare blocks exist so quit\n");
+               ret = FAIL;
+       }
+
+       return ret;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Static_Wear_Leveling
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=1
+* Description:  This is static wear leveling (done by explicit call)
+*               search for most&least used
+*               if difference < GATE:
+*                   update the block table with exhange
+*                   mark block table in flash as IN_PROGRESS
+*                   copy flash block
+*               the caller should handle GC clean up after calling this function
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int FTL_Static_Wear_Leveling(void)
+{
+       u8 most_worn_cnt;
+       u8 least_worn_cnt;
+       u32 most_worn_idx;
+       u32 least_worn_idx;
+       int result = PASS;
+       int go_on = PASS;
+       u32 replaced_blks = 0;
+       u8 *chang_flag = flags_static_wear_leveling;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (!chang_flag)
+               return FAIL;
+
+       memset(chang_flag, FAIL, DeviceInfo.wDataBlockNum);
+       while (go_on == PASS) {
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "starting static wear leveling\n");
+               most_worn_cnt = 0;
+               least_worn_cnt = 0xFF;
+               least_worn_idx = BLOCK_TABLE_INDEX;
+               most_worn_idx = BLOCK_TABLE_INDEX;
+
+               find_least_most_worn(chang_flag, &least_worn_idx,
+                       &least_worn_cnt, &most_worn_idx, &most_worn_cnt);
+
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Used and least worn is block %u, whos count is %u\n",
+                       (unsigned int)least_worn_idx,
+                       (unsigned int)least_worn_cnt);
+
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Free and  most worn is block %u, whos count is %u\n",
+                       (unsigned int)most_worn_idx,
+                       (unsigned int)most_worn_cnt);
+
+               if ((most_worn_cnt > least_worn_cnt) &&
+                       (most_worn_cnt - least_worn_cnt > WEAR_LEVELING_GATE))
+                       go_on = move_blks_for_wear_leveling(chang_flag,
+                               &least_worn_idx, &replaced_blks, &result);
+               else
+                       go_on = FAIL;
+       }
+
+       return result;
+}
+
+#if CMD_DMA
+static int do_garbage_collection(u32 discard_cnt)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 pba;
+       u8 bt_block_erased = 0;
+       int i, cnt, ret = FAIL;
+       u64 addr;
+
+       i = 0;
+       while ((i < DeviceInfo.wDataBlockNum) && (discard_cnt > 0) &&
+                       ((ftl_cmd_cnt + 28) < 256)) {
+               if (((pbt[i] & BAD_BLOCK) != BAD_BLOCK) &&
+                               (pbt[i] & DISCARD_BLOCK)) {
+                       if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+                               g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                               FTL_Write_IN_Progress_Block_Table_Page();
+                       }
+
+                       addr = FTL_Get_Physical_Block_Addr((u64)i *
+                                               DeviceInfo.wBlockDataSize);
+                       pba = BLK_FROM_ADDR(addr);
+
+                       for (cnt = FIRST_BT_ID; cnt <= LAST_BT_ID; cnt++) {
+                               if (pba == g_pBTBlocks[cnt - FIRST_BT_ID]) {
+                                       nand_dbg_print(NAND_DBG_DEBUG,
+                                               "GC will erase BT block %u\n",
+                                               (unsigned int)pba);
+                                       discard_cnt--;
+                                       i++;
+                                       bt_block_erased = 1;
+                                       break;
+                               }
+                       }
+
+                       if (bt_block_erased) {
+                               bt_block_erased = 0;
+                               continue;
+                       }
+
+                       addr = FTL_Get_Physical_Block_Addr((u64)i *
+                                               DeviceInfo.wBlockDataSize);
+
+                       if (PASS == GLOB_FTL_Block_Erase(addr)) {
+                               pbt[i] &= (u32)(~DISCARD_BLOCK);
+                               pbt[i] |= (u32)(SPARE_BLOCK);
+                               p_BTableChangesDelta =
+                                       (struct BTableChangesDelta *)
+                                       g_pBTDelta_Free;
+                               g_pBTDelta_Free +=
+                                       sizeof(struct BTableChangesDelta);
+                               p_BTableChangesDelta->ftl_cmd_cnt =
+                                       ftl_cmd_cnt - 1;
+                               p_BTableChangesDelta->BT_Index = i;
+                               p_BTableChangesDelta->BT_Entry_Value = pbt[i];
+                               p_BTableChangesDelta->ValidFields = 0x0C;
+                               discard_cnt--;
+                               ret = PASS;
+                       } else {
+                               MARK_BLOCK_AS_BAD(pbt[i]);
+                       }
+               }
+
+               i++;
+       }
+
+       return ret;
+}
+
+#else
+static int do_garbage_collection(u32 discard_cnt)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 pba;
+       u8 bt_block_erased = 0;
+       int i, cnt, ret = FAIL;
+       u64 addr;
+
+       i = 0;
+       while ((i < DeviceInfo.wDataBlockNum) && (discard_cnt > 0)) {
+               if (((pbt[i] & BAD_BLOCK) != BAD_BLOCK) &&
+                               (pbt[i] & DISCARD_BLOCK)) {
+                       if (IN_PROGRESS_BLOCK_TABLE != g_cBlockTableStatus) {
+                               g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                               FTL_Write_IN_Progress_Block_Table_Page();
+                       }
+
+                       addr = FTL_Get_Physical_Block_Addr((u64)i *
+                                               DeviceInfo.wBlockDataSize);
+                       pba = BLK_FROM_ADDR(addr);
+
+                       for (cnt = FIRST_BT_ID; cnt <= LAST_BT_ID; cnt++) {
+                               if (pba == g_pBTBlocks[cnt - FIRST_BT_ID]) {
+                                       nand_dbg_print(NAND_DBG_DEBUG,
+                                               "GC will erase BT block %d\n",
+                                               pba);
+                                       discard_cnt--;
+                                       i++;
+                                       bt_block_erased = 1;
+                                       break;
+                               }
+                       }
+
+                       if (bt_block_erased) {
+                               bt_block_erased = 0;
+                               continue;
+                       }
+
+                       /* If the discard block is L2 cache block, then just skip it */
+                       for (cnt = 0; cnt < BLK_NUM_FOR_L2_CACHE; cnt++) {
+                               if (cache_l2.blk_array[cnt] == pba) {
+                                       nand_dbg_print(NAND_DBG_DEBUG,
+                                               "GC will erase L2 cache blk %d\n",
+                                               pba);
+                                       break;
+                               }
+                       }
+                       if (cnt < BLK_NUM_FOR_L2_CACHE) { /* Skip it */
+                               discard_cnt--;
+                               i++;
+                               continue;
+                       }
+
+                       addr = FTL_Get_Physical_Block_Addr((u64)i *
+                                               DeviceInfo.wBlockDataSize);
+
+                       if (PASS == GLOB_FTL_Block_Erase(addr)) {
+                               pbt[i] &= (u32)(~DISCARD_BLOCK);
+                               pbt[i] |= (u32)(SPARE_BLOCK);
+                               discard_cnt--;
+                               ret = PASS;
+                       } else {
+                               MARK_BLOCK_AS_BAD(pbt[i]);
+                       }
+               }
+
+               i++;
+       }
+
+       return ret;
+}
+#endif
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Garbage_Collection
+* Inputs:       none
+* Outputs:      PASS / FAIL (returns the number of un-erased blocks
+* Description:  search the block table for all discarded blocks to erase
+*               for each discarded block:
+*                   set the flash block to IN_PROGRESS
+*                   erase the block
+*                   update the block table
+*                   write the block table to flash
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Garbage_Collection(void)
+{
+       u32 i;
+       u32 wDiscard = 0;
+       int wResult = FAIL;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       if (GC_Called) {
+               printk(KERN_ALERT "GLOB_FTL_Garbage_Collection() "
+                       "has been re-entered! Exit.\n");
+               return PASS;
+       }
+
+       GC_Called = 1;
+
+       GLOB_FTL_BT_Garbage_Collection();
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_DISCARDED_BLOCK(i))
+                       wDiscard++;
+       }
+
+       if (wDiscard <= 0) {
+               GC_Called = 0;
+               return wResult;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "Found %d discarded blocks\n", wDiscard);
+
+       FTL_Write_Block_Table(FAIL);
+
+       wResult = do_garbage_collection(wDiscard);
+
+       FTL_Write_Block_Table(FAIL);
+
+       GC_Called = 0;
+
+       return wResult;
+}
+
+
+#if CMD_DMA
+static int do_bt_garbage_collection(void)
+{
+       u32 pba, lba;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 *pBTBlocksNode = (u32 *)g_pBTBlocks;
+       u64 addr;
+       int i, ret = FAIL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       if (BT_GC_Called)
+               return PASS;
+
+       BT_GC_Called = 1;
+
+       for (i = last_erased; (i <= LAST_BT_ID) &&
+               (g_pBTBlocks[((i + 2) % (1 + LAST_BT_ID - FIRST_BT_ID)) +
+               FIRST_BT_ID - FIRST_BT_ID] != BTBLOCK_INVAL) &&
+               ((ftl_cmd_cnt + 28)) < 256; i++) {
+               pba = pBTBlocksNode[i - FIRST_BT_ID];
+               lba = FTL_Get_Block_Index(pba);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "do_bt_garbage_collection: pba %d, lba %d\n",
+                       pba, lba);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Block Table Entry: %d", pbt[lba]);
+
+               if (((pbt[lba] & BAD_BLOCK) != BAD_BLOCK) &&
+                       (pbt[lba] & DISCARD_BLOCK)) {
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "do_bt_garbage_collection_cdma: "
+                               "Erasing Block tables present in block %d\n",
+                               pba);
+                       addr = FTL_Get_Physical_Block_Addr((u64)lba *
+                                               DeviceInfo.wBlockDataSize);
+                       if (PASS == GLOB_FTL_Block_Erase(addr)) {
+                               pbt[lba] &= (u32)(~DISCARD_BLOCK);
+                               pbt[lba] |= (u32)(SPARE_BLOCK);
+
+                               p_BTableChangesDelta =
+                                       (struct BTableChangesDelta *)
+                                       g_pBTDelta_Free;
+                               g_pBTDelta_Free +=
+                                       sizeof(struct BTableChangesDelta);
+
+                               p_BTableChangesDelta->ftl_cmd_cnt =
+                                       ftl_cmd_cnt - 1;
+                               p_BTableChangesDelta->BT_Index = lba;
+                               p_BTableChangesDelta->BT_Entry_Value =
+                                                               pbt[lba];
+
+                               p_BTableChangesDelta->ValidFields = 0x0C;
+
+                               ret = PASS;
+                               pBTBlocksNode[last_erased - FIRST_BT_ID] =
+                                                       BTBLOCK_INVAL;
+                               nand_dbg_print(NAND_DBG_DEBUG,
+                                       "resetting bt entry at index %d "
+                                       "value %d\n", i,
+                                       pBTBlocksNode[i - FIRST_BT_ID]);
+                               if (last_erased == LAST_BT_ID)
+                                       last_erased = FIRST_BT_ID;
+                               else
+                                       last_erased++;
+                       } else {
+                               MARK_BLOCK_AS_BAD(pbt[lba]);
+                       }
+               }
+       }
+
+       BT_GC_Called = 0;
+
+       return ret;
+}
+
+#else
+static int do_bt_garbage_collection(void)
+{
+       u32 pba, lba;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 *pBTBlocksNode = (u32 *)g_pBTBlocks;
+       u64 addr;
+       int i, ret = FAIL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       if (BT_GC_Called)
+               return PASS;
+
+       BT_GC_Called = 1;
+
+       for (i = last_erased; (i <= LAST_BT_ID) &&
+               (g_pBTBlocks[((i + 2) % (1 + LAST_BT_ID - FIRST_BT_ID)) +
+               FIRST_BT_ID - FIRST_BT_ID] != BTBLOCK_INVAL); i++) {
+               pba = pBTBlocksNode[i - FIRST_BT_ID];
+               lba = FTL_Get_Block_Index(pba);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "do_bt_garbage_collection_cdma: pba %d, lba %d\n",
+                       pba, lba);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Block Table Entry: %d", pbt[lba]);
+
+               if (((pbt[lba] & BAD_BLOCK) != BAD_BLOCK) &&
+                       (pbt[lba] & DISCARD_BLOCK)) {
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "do_bt_garbage_collection: "
+                               "Erasing Block tables present in block %d\n",
+                               pba);
+                       addr = FTL_Get_Physical_Block_Addr((u64)lba *
+                                               DeviceInfo.wBlockDataSize);
+                       if (PASS == GLOB_FTL_Block_Erase(addr)) {
+                               pbt[lba] &= (u32)(~DISCARD_BLOCK);
+                               pbt[lba] |= (u32)(SPARE_BLOCK);
+                               ret = PASS;
+                               pBTBlocksNode[last_erased - FIRST_BT_ID] =
+                                                       BTBLOCK_INVAL;
+                               nand_dbg_print(NAND_DBG_DEBUG,
+                                       "resetting bt entry at index %d "
+                                       "value %d\n", i,
+                                       pBTBlocksNode[i - FIRST_BT_ID]);
+                               if (last_erased == LAST_BT_ID)
+                                       last_erased = FIRST_BT_ID;
+                               else
+                                       last_erased++;
+                       } else {
+                               MARK_BLOCK_AS_BAD(pbt[lba]);
+                       }
+               }
+       }
+
+       BT_GC_Called = 0;
+
+       return ret;
+}
+
+#endif
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_BT_Garbage_Collection
+* Inputs:       none
+* Outputs:      PASS / FAIL (returns the number of un-erased blocks
+* Description:  Erases discarded blocks containing Block table
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_BT_Garbage_Collection(void)
+{
+       return do_bt_garbage_collection();
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Replace_OneBlock
+* Inputs:       Block number 1
+*               Block number 2
+* Outputs:      Replaced Block Number
+* Description:  Interchange block table entries at wBlockNum and wReplaceNum
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Replace_OneBlock(u32 blk, u32 rep_blk)
+{
+       u32 tmp_blk;
+       u32 replace_node = BAD_BLOCK;
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       if (rep_blk != BAD_BLOCK) {
+               if (IS_BAD_BLOCK(blk))
+                       tmp_blk = pbt[blk];
+               else
+                       tmp_blk = DISCARD_BLOCK | (~SPARE_BLOCK & pbt[blk]);
+
+               replace_node = (u32) ((~SPARE_BLOCK) & pbt[rep_blk]);
+               pbt[blk] = replace_node;
+               pbt[rep_blk] = tmp_blk;
+
+#if CMD_DMA
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+               p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+               p_BTableChangesDelta->BT_Index = blk;
+               p_BTableChangesDelta->BT_Entry_Value = pbt[blk];
+
+               p_BTableChangesDelta->ValidFields = 0x0C;
+
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+               p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+               p_BTableChangesDelta->BT_Index = rep_blk;
+               p_BTableChangesDelta->BT_Entry_Value = pbt[rep_blk];
+               p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+       }
+
+       return replace_node;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Write_Block_Table_Data
+* Inputs:       Block table size in pages
+* Outputs:      PASS=0 / FAIL=1
+* Description:  Write block table data in flash
+*               If first page and last page
+*                  Write data+BT flag
+*               else
+*                  Write data
+*               BT flag is a counter. Its value is incremented for block table
+*               write in a new Block
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Write_Block_Table_Data(void)
+{
+       u64 dwBlockTableAddr, pTempAddr;
+       u32 Block;
+       u16 Page, PageCount;
+       u8 *tempBuf = tmp_buf_write_blk_table_data;
+       int wBytesCopied;
+       u16 bt_pages;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       dwBlockTableAddr =
+               (u64)((u64)g_wBlockTableIndex * DeviceInfo.wBlockDataSize +
+               (u64)g_wBlockTableOffset * DeviceInfo.wPageDataSize);
+       pTempAddr = dwBlockTableAddr;
+
+       bt_pages = FTL_Get_Block_Table_Flash_Size_Pages();
+
+       nand_dbg_print(NAND_DBG_DEBUG, "FTL_Write_Block_Table_Data: "
+                              "page= %d BlockTableIndex= %d "
+                              "BlockTableOffset=%d\n", bt_pages,
+                              g_wBlockTableIndex, g_wBlockTableOffset);
+
+       Block = BLK_FROM_ADDR(pTempAddr);
+       Page = PAGE_FROM_ADDR(pTempAddr, Block);
+       PageCount = 1;
+
+       if (bt_block_changed) {
+               if (bt_flag == LAST_BT_ID) {
+                       bt_flag = FIRST_BT_ID;
+                       g_pBTBlocks[bt_flag - FIRST_BT_ID] = Block;
+               } else if (bt_flag < LAST_BT_ID) {
+                       bt_flag++;
+                       g_pBTBlocks[bt_flag - FIRST_BT_ID] = Block;
+               }
+
+               if ((bt_flag > (LAST_BT_ID-4)) &&
+                       g_pBTBlocks[FIRST_BT_ID - FIRST_BT_ID] !=
+                                               BTBLOCK_INVAL) {
+                       bt_block_changed = 0;
+                       GLOB_FTL_BT_Garbage_Collection();
+               }
+
+               bt_block_changed = 0;
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Block Table Counter is %u Block %u\n",
+                       bt_flag, (unsigned int)Block);
+       }
+
+       memset(tempBuf, 0, 3);
+       tempBuf[3] = bt_flag;
+       wBytesCopied = FTL_Copy_Block_Table_To_Flash(tempBuf + 4,
+                       DeviceInfo.wPageDataSize - 4, 0);
+       memset(&tempBuf[wBytesCopied + 4], 0xff,
+               DeviceInfo.wPageSize - (wBytesCopied + 4));
+       FTL_Insert_Block_Table_Signature(&tempBuf[DeviceInfo.wPageDataSize],
+                                       bt_flag);
+
+#if CMD_DMA
+       memcpy(g_pNextBlockTable, tempBuf,
+               DeviceInfo.wPageSize * sizeof(u8));
+       nand_dbg_print(NAND_DBG_DEBUG, "Writing First Page of Block Table "
+               "Block %u Page %u\n", (unsigned int)Block, Page);
+       if (FAIL == GLOB_LLD_Write_Page_Main_Spare_cdma(g_pNextBlockTable,
+               Block, Page, 1,
+               LLD_CMD_FLAG_MODE_CDMA | LLD_CMD_FLAG_ORDER_BEFORE_REST)) {
+               nand_dbg_print(NAND_DBG_WARN, "NAND Program fail in "
+                       "%s, Line %d, Function: %s, "
+                       "new Bad Block %d generated!\n",
+                       __FILE__, __LINE__, __func__, Block);
+               goto func_return;
+       }
+
+       ftl_cmd_cnt++;
+       g_pNextBlockTable += ((DeviceInfo.wPageSize * sizeof(u8)));
+#else
+       if (FAIL == GLOB_LLD_Write_Page_Main_Spare(tempBuf, Block, Page, 1)) {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "NAND Program fail in %s, Line %d, Function: %s, "
+                       "new Bad Block %d generated!\n",
+                       __FILE__, __LINE__, __func__, Block);
+               goto func_return;
+       }
+#endif
+
+       if (bt_pages > 1) {
+               PageCount = bt_pages - 1;
+               if (PageCount > 1) {
+                       wBytesCopied += FTL_Copy_Block_Table_To_Flash(tempBuf,
+                               DeviceInfo.wPageDataSize * (PageCount - 1),
+                               wBytesCopied);
+
+#if CMD_DMA
+                       memcpy(g_pNextBlockTable, tempBuf,
+                               (PageCount - 1) * DeviceInfo.wPageDataSize);
+                       if (FAIL == GLOB_LLD_Write_Page_Main_cdma(
+                               g_pNextBlockTable, Block, Page + 1,
+                               PageCount - 1)) {
+                               nand_dbg_print(NAND_DBG_WARN,
+                                       "NAND Program fail in %s, Line %d, "
+                                       "Function: %s, "
+                                       "new Bad Block %d generated!\n",
+                                       __FILE__, __LINE__, __func__,
+                                       (int)Block);
+                               goto func_return;
+                       }
+
+                       ftl_cmd_cnt++;
+                       g_pNextBlockTable += (PageCount - 1) *
+                               DeviceInfo.wPageDataSize * sizeof(u8);
+#else
+                       if (FAIL == GLOB_LLD_Write_Page_Main(tempBuf,
+                                       Block, Page + 1, PageCount - 1)) {
+                               nand_dbg_print(NAND_DBG_WARN,
+                                       "NAND Program fail in %s, Line %d, "
+                                       "Function: %s, "
+                                       "new Bad Block %d generated!\n",
+                                       __FILE__, __LINE__, __func__,
+                                       (int)Block);
+                               goto func_return;
+                       }
+#endif
+               }
+
+               wBytesCopied = FTL_Copy_Block_Table_To_Flash(tempBuf,
+                               DeviceInfo.wPageDataSize, wBytesCopied);
+               memset(&tempBuf[wBytesCopied], 0xff,
+                       DeviceInfo.wPageSize-wBytesCopied);
+               FTL_Insert_Block_Table_Signature(
+                       &tempBuf[DeviceInfo.wPageDataSize], bt_flag);
+#if CMD_DMA
+               memcpy(g_pNextBlockTable, tempBuf,
+                               DeviceInfo.wPageSize * sizeof(u8));
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Writing the last Page of Block Table "
+                       "Block %u Page %u\n",
+                       (unsigned int)Block, Page + bt_pages - 1);
+               if (FAIL == GLOB_LLD_Write_Page_Main_Spare_cdma(
+                       g_pNextBlockTable, Block, Page + bt_pages - 1, 1,
+                       LLD_CMD_FLAG_MODE_CDMA |
+                       LLD_CMD_FLAG_ORDER_BEFORE_REST)) {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "NAND Program fail in %s, Line %d, "
+                               "Function: %s, new Bad Block %d generated!\n",
+                               __FILE__, __LINE__, __func__, Block);
+                       goto func_return;
+               }
+               ftl_cmd_cnt++;
+#else
+               if (FAIL == GLOB_LLD_Write_Page_Main_Spare(tempBuf,
+                                       Block, Page+bt_pages - 1, 1)) {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "NAND Program fail in %s, Line %d, "
+                               "Function: %s, "
+                               "new Bad Block %d generated!\n",
+                               __FILE__, __LINE__, __func__, Block);
+                       goto func_return;
+               }
+#endif
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "FTL_Write_Block_Table_Data: done\n");
+
+func_return:
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Replace_Block_Table
+* Inputs:       None
+* Outputs:      PASS=0 / FAIL=1
+* Description:  Get a new block to write block table
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Replace_Block_Table(void)
+{
+       u32 blk;
+       int gc;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       blk = FTL_Replace_LWBlock(BLOCK_TABLE_INDEX, &gc);
+
+       if ((BAD_BLOCK == blk) && (PASS == gc)) {
+               GLOB_FTL_Garbage_Collection();
+               blk = FTL_Replace_LWBlock(BLOCK_TABLE_INDEX, &gc);
+       }
+       if (BAD_BLOCK == blk)
+               printk(KERN_ERR "%s, %s: There is no spare block. "
+                       "It should never happen\n",
+                       __FILE__, __func__);
+
+       nand_dbg_print(NAND_DBG_DEBUG, "New Block table Block is %d\n", blk);
+
+       return blk;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Replace_LWBlock
+* Inputs:       Block number
+*               Pointer to Garbage Collect flag
+* Outputs:
+* Description:  Determine the least weared block by traversing
+*               block table
+*               Set Garbage collection to be called if number of spare
+*               block is less than Free Block Gate count
+*               Change Block table entry to map least worn block for current
+*               operation
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Replace_LWBlock(u32 wBlockNum, int *pGarbageCollect)
+{
+       u32 i;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u8 wLeastWornCounter = 0xFF;
+       u32 wLeastWornIndex = BAD_BLOCK;
+       u32 wSpareBlockNum = 0;
+       u32 wDiscardBlockNum = 0;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       if (IS_SPARE_BLOCK(wBlockNum)) {
+               *pGarbageCollect = FAIL;
+               pbt[wBlockNum] = (u32)(pbt[wBlockNum] & (~SPARE_BLOCK));
+#if CMD_DMA
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+               p_BTableChangesDelta->ftl_cmd_cnt =
+                                               ftl_cmd_cnt;
+               p_BTableChangesDelta->BT_Index = (u32)(wBlockNum);
+               p_BTableChangesDelta->BT_Entry_Value = pbt[wBlockNum];
+               p_BTableChangesDelta->ValidFields = 0x0C;
+#endif
+               return pbt[wBlockNum];
+       }
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_DISCARDED_BLOCK(i))
+                       wDiscardBlockNum++;
+
+               if (IS_SPARE_BLOCK(i)) {
+                       u32 wPhysicalIndex = (u32)((~BAD_BLOCK) & pbt[i]);
+                       if (wPhysicalIndex > DeviceInfo.wSpectraEndBlock)
+                               printk(KERN_ERR "FTL_Replace_LWBlock: "
+                                       "This should never occur!\n");
+                       if (g_pWearCounter[wPhysicalIndex -
+                               DeviceInfo.wSpectraStartBlock] <
+                               wLeastWornCounter) {
+                               wLeastWornCounter =
+                                       g_pWearCounter[wPhysicalIndex -
+                                       DeviceInfo.wSpectraStartBlock];
+                               wLeastWornIndex = i;
+                       }
+                       wSpareBlockNum++;
+               }
+       }
+
+       nand_dbg_print(NAND_DBG_WARN,
+               "FTL_Replace_LWBlock: Least Worn Counter %d\n",
+               (int)wLeastWornCounter);
+
+       if ((wDiscardBlockNum >= NUM_FREE_BLOCKS_GATE) ||
+               (wSpareBlockNum <= NUM_FREE_BLOCKS_GATE))
+               *pGarbageCollect = PASS;
+       else
+               *pGarbageCollect = FAIL;
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "FTL_Replace_LWBlock: Discarded Blocks %u Spare"
+               " Blocks %u\n",
+               (unsigned int)wDiscardBlockNum,
+               (unsigned int)wSpareBlockNum);
+
+       return FTL_Replace_OneBlock(wBlockNum, wLeastWornIndex);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Replace_MWBlock
+* Inputs:       None
+* Outputs:      most worn spare block no./BAD_BLOCK
+* Description:  It finds most worn spare block.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static u32 FTL_Replace_MWBlock(void)
+{
+       u32 i;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u8 wMostWornCounter = 0;
+       u32 wMostWornIndex = BAD_BLOCK;
+       u32 wSpareBlockNum = 0;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_SPARE_BLOCK(i)) {
+                       u32 wPhysicalIndex = (u32)((~SPARE_BLOCK) & pbt[i]);
+                       if (g_pWearCounter[wPhysicalIndex -
+                           DeviceInfo.wSpectraStartBlock] >
+                           wMostWornCounter) {
+                               wMostWornCounter =
+                                   g_pWearCounter[wPhysicalIndex -
+                                   DeviceInfo.wSpectraStartBlock];
+                               wMostWornIndex = wPhysicalIndex;
+                       }
+                       wSpareBlockNum++;
+               }
+       }
+
+       if (wSpareBlockNum <= 2)
+               return BAD_BLOCK;
+
+       return wMostWornIndex;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Replace_Block
+* Inputs:       Block Address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  If block specified by blk_addr parameter is not free,
+*               replace it with the least worn block.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Replace_Block(u64 blk_addr)
+{
+       u32 current_blk = BLK_FROM_ADDR(blk_addr);
+       u32 *pbt = (u32 *)g_pBlockTable;
+       int wResult = PASS;
+       int GarbageCollect = FAIL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       if (IS_SPARE_BLOCK(current_blk)) {
+               pbt[current_blk] = (~SPARE_BLOCK) & pbt[current_blk];
+#if CMD_DMA
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+               p_BTableChangesDelta->ftl_cmd_cnt =
+                       ftl_cmd_cnt;
+               p_BTableChangesDelta->BT_Index = current_blk;
+               p_BTableChangesDelta->BT_Entry_Value = pbt[current_blk];
+               p_BTableChangesDelta->ValidFields = 0x0C ;
+#endif
+               return wResult;
+       }
+
+       FTL_Replace_LWBlock(current_blk, &GarbageCollect);
+
+       if (PASS == GarbageCollect)
+               wResult = GLOB_FTL_Garbage_Collection();
+
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Is_BadBlock
+* Inputs:       block number to test
+* Outputs:      PASS (block is BAD) / FAIL (block is not bad)
+* Description:  test if this block number is flagged as bad
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Is_BadBlock(u32 wBlockNum)
+{
+       u32 *pbt = (u32 *)g_pBlockTable;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       if (wBlockNum >= DeviceInfo.wSpectraStartBlock
+               && BAD_BLOCK == (pbt[wBlockNum] & BAD_BLOCK))
+               return PASS;
+       else
+               return FAIL;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Flush_Cache
+* Inputs:       none
+* Outputs:      PASS=0 / FAIL=1
+* Description:  flush all the cache blocks to flash
+*               if a cache block is not dirty, don't do anything with it
+*               else, write the block and update the block table
+* Note:         This function should be called at shutdown/power down.
+*               to write important data into device
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Flush_Cache(void)
+{
+       int i, ret;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < CACHE_ITEM_NUM; i++) {
+               if (SET == Cache.array[i].changed) {
+#if CMD_DMA
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+                       int_cache[ftl_cmd_cnt].item = i;
+                       int_cache[ftl_cmd_cnt].cache.address =
+                                       Cache.array[i].address;
+                       int_cache[ftl_cmd_cnt].cache.changed = CLEAR;
+#endif
+#endif
+                       ret = write_back_to_l2_cache(Cache.array[i].buf, Cache.array[i].address);
+                       if (PASS == ret) {
+                               Cache.array[i].changed = CLEAR;
+                       } else {
+                               printk(KERN_ALERT "Failed when write back to L2 cache!\n");
+                               /* TODO - How to handle this? */
+                       }
+               }
+       }
+
+       flush_l2_cache();
+
+       return FTL_Write_Block_Table(FAIL);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Page_Read
+* Inputs:       pointer to data
+*                   logical address of data (u64 is LBA * Bytes/Page)
+* Outputs:      PASS=0 / FAIL=1
+* Description:  reads a page of data into RAM from the cache
+*               if the data is not already in cache, read from flash to cache
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Page_Read(u8 *data, u64 logical_addr)
+{
+       u16 cache_item;
+       int res = PASS;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "GLOB_FTL_Page_Read - "
+               "page_addr: %llu\n", logical_addr);
+
+       cache_item = FTL_Cache_If_Hit(logical_addr);
+
+       if (UNHIT_CACHE_ITEM == cache_item) {
+               nand_dbg_print(NAND_DBG_DEBUG,
+                              "GLOB_FTL_Page_Read: Cache not hit\n");
+               res = FTL_Cache_Write();
+               if (ERR == FTL_Cache_Read(logical_addr))
+                       res = ERR;
+               cache_item = Cache.LRU;
+       }
+
+       FTL_Cache_Read_Page(data, logical_addr, cache_item);
+
+       return res;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Page_Write
+* Inputs:       pointer to data
+*               address of data (ADDRESSTYPE is LBA * Bytes/Page)
+* Outputs:      PASS=0 / FAIL=1
+* Description:  writes a page of data from RAM to the cache
+*               if the data is not already in cache, write back the
+*               least recently used block and read the addressed block
+*               from flash to cache
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Page_Write(u8 *pData, u64 dwPageAddr)
+{
+       u16 cache_blk;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       int wResult = PASS;
+
+       nand_dbg_print(NAND_DBG_TRACE, "GLOB_FTL_Page_Write - "
+               "dwPageAddr: %llu\n", dwPageAddr);
+
+       cache_blk = FTL_Cache_If_Hit(dwPageAddr);
+
+       if (UNHIT_CACHE_ITEM == cache_blk) {
+               wResult = FTL_Cache_Write();
+               if (IS_BAD_BLOCK(BLK_FROM_ADDR(dwPageAddr))) {
+                       wResult = FTL_Replace_Block(dwPageAddr);
+                       pbt[BLK_FROM_ADDR(dwPageAddr)] |= SPARE_BLOCK;
+                       if (wResult == FAIL)
+                               return FAIL;
+               }
+               if (ERR == FTL_Cache_Read(dwPageAddr))
+                       wResult = ERR;
+               cache_blk = Cache.LRU;
+               FTL_Cache_Write_Page(pData, dwPageAddr, cache_blk, 0);
+       } else {
+#if CMD_DMA
+               FTL_Cache_Write_Page(pData, dwPageAddr, cache_blk,
+                               LLD_CMD_FLAG_ORDER_BEFORE_REST);
+#else
+               FTL_Cache_Write_Page(pData, dwPageAddr, cache_blk, 0);
+#endif
+       }
+
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     GLOB_FTL_Block_Erase
+* Inputs:       address of block to erase (now in byte format, should change to
+* block format)
+* Outputs:      PASS=0 / FAIL=1
+* Description:  erases the specified block
+*               increments the erase count
+*               If erase count reaches its upper limit,call function to
+*               do the ajustment as per the relative erase count values
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int GLOB_FTL_Block_Erase(u64 blk_addr)
+{
+       int status;
+       u32 BlkIdx;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       BlkIdx = (u32)(blk_addr >> DeviceInfo.nBitsInBlockDataSize);
+
+       if (BlkIdx < DeviceInfo.wSpectraStartBlock) {
+               printk(KERN_ERR "GLOB_FTL_Block_Erase: "
+                       "This should never occur\n");
+               return FAIL;
+       }
+
+#if CMD_DMA
+       status = GLOB_LLD_Erase_Block_cdma(BlkIdx, LLD_CMD_FLAG_MODE_CDMA);
+       if (status == FAIL)
+               nand_dbg_print(NAND_DBG_WARN,
+                              "NAND Program fail in %s, Line %d, "
+                              "Function: %s, new Bad Block %d generated!\n",
+                              __FILE__, __LINE__, __func__, BlkIdx);
+#else
+       status = GLOB_LLD_Erase_Block(BlkIdx);
+       if (status == FAIL) {
+               nand_dbg_print(NAND_DBG_WARN,
+                              "NAND Program fail in %s, Line %d, "
+                              "Function: %s, new Bad Block %d generated!\n",
+                              __FILE__, __LINE__, __func__, BlkIdx);
+               return status;
+       }
+#endif
+
+       if (DeviceInfo.MLCDevice) {
+               g_pReadCounter[BlkIdx - DeviceInfo.wSpectraStartBlock] = 0;
+               if (g_cBlockTableStatus != IN_PROGRESS_BLOCK_TABLE) {
+                       g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                       FTL_Write_IN_Progress_Block_Table_Page();
+               }
+       }
+
+       g_pWearCounter[BlkIdx - DeviceInfo.wSpectraStartBlock]++;
+
+#if CMD_DMA
+       p_BTableChangesDelta =
+               (struct BTableChangesDelta *)g_pBTDelta_Free;
+       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+       p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+       p_BTableChangesDelta->WC_Index =
+               BlkIdx - DeviceInfo.wSpectraStartBlock;
+       p_BTableChangesDelta->WC_Entry_Value =
+               g_pWearCounter[BlkIdx - DeviceInfo.wSpectraStartBlock];
+       p_BTableChangesDelta->ValidFields = 0x30;
+
+       if (DeviceInfo.MLCDevice) {
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+               p_BTableChangesDelta->ftl_cmd_cnt =
+                       ftl_cmd_cnt;
+               p_BTableChangesDelta->RC_Index =
+                       BlkIdx - DeviceInfo.wSpectraStartBlock;
+               p_BTableChangesDelta->RC_Entry_Value =
+                       g_pReadCounter[BlkIdx -
+                               DeviceInfo.wSpectraStartBlock];
+               p_BTableChangesDelta->ValidFields = 0xC0;
+       }
+
+       ftl_cmd_cnt++;
+#endif
+
+       if (g_pWearCounter[BlkIdx - DeviceInfo.wSpectraStartBlock] == 0xFE)
+               FTL_Adjust_Relative_Erase_Count(BlkIdx);
+
+       return status;
+}
+
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Adjust_Relative_Erase_Count
+* Inputs:       index to block that was just incremented and is at the max
+* Outputs:      PASS=0 / FAIL=1
+* Description:  If any erase counts at MAX, adjusts erase count of every
+*               block by substracting least worn
+*               counter from counter value of every entry in wear table
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+static int FTL_Adjust_Relative_Erase_Count(u32 Index_of_MAX)
+{
+       u8 wLeastWornCounter = MAX_BYTE_VALUE;
+       u8 wWearCounter;
+       u32 i, wWearIndex;
+       u32 *pbt = (u32 *)g_pBlockTable;
+       int wResult = PASS;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+               __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < DeviceInfo.wDataBlockNum; i++) {
+               if (IS_BAD_BLOCK(i))
+                       continue;
+               wWearIndex = (u32)(pbt[i] & (~BAD_BLOCK));
+
+               if ((wWearIndex - DeviceInfo.wSpectraStartBlock) < 0)
+                       printk(KERN_ERR "FTL_Adjust_Relative_Erase_Count:"
+                                       "This should never occur\n");
+               wWearCounter = g_pWearCounter[wWearIndex -
+                       DeviceInfo.wSpectraStartBlock];
+               if (wWearCounter < wLeastWornCounter)
+                       wLeastWornCounter = wWearCounter;
+       }
+
+       if (wLeastWornCounter == 0) {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Adjusting Wear Levelling Counters: Special Case\n");
+               g_pWearCounter[Index_of_MAX -
+                       DeviceInfo.wSpectraStartBlock]--;
+#if CMD_DMA
+               p_BTableChangesDelta =
+                       (struct BTableChangesDelta *)g_pBTDelta_Free;
+               g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+               p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+               p_BTableChangesDelta->WC_Index =
+                       Index_of_MAX - DeviceInfo.wSpectraStartBlock;
+               p_BTableChangesDelta->WC_Entry_Value =
+                       g_pWearCounter[Index_of_MAX -
+                               DeviceInfo.wSpectraStartBlock];
+               p_BTableChangesDelta->ValidFields = 0x30;
+#endif
+               FTL_Static_Wear_Leveling();
+       } else {
+               for (i = 0; i < DeviceInfo.wDataBlockNum; i++)
+                       if (!IS_BAD_BLOCK(i)) {
+                               wWearIndex = (u32)(pbt[i] & (~BAD_BLOCK));
+                               g_pWearCounter[wWearIndex -
+                                       DeviceInfo.wSpectraStartBlock] =
+                                       (u8)(g_pWearCounter
+                                       [wWearIndex -
+                                       DeviceInfo.wSpectraStartBlock] -
+                                       wLeastWornCounter);
+#if CMD_DMA
+                               p_BTableChangesDelta =
+                               (struct BTableChangesDelta *)g_pBTDelta_Free;
+                               g_pBTDelta_Free +=
+                                       sizeof(struct BTableChangesDelta);
+
+                               p_BTableChangesDelta->ftl_cmd_cnt =
+                                       ftl_cmd_cnt;
+                               p_BTableChangesDelta->WC_Index = wWearIndex -
+                                       DeviceInfo.wSpectraStartBlock;
+                               p_BTableChangesDelta->WC_Entry_Value =
+                                       g_pWearCounter[wWearIndex -
+                                       DeviceInfo.wSpectraStartBlock];
+                               p_BTableChangesDelta->ValidFields = 0x30;
+#endif
+                       }
+       }
+
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Write_IN_Progress_Block_Table_Page
+* Inputs:       None
+* Outputs:      None
+* Description:  It writes in-progress flag page to the page next to
+*               block table
+***********************************************************************/
+static int FTL_Write_IN_Progress_Block_Table_Page(void)
+{
+       int wResult = PASS;
+       u16 bt_pages;
+       u16 dwIPFPageAddr;
+#if CMD_DMA
+#else
+       u32 *pbt = (u32 *)g_pBlockTable;
+       u32 wTempBlockTableIndex;
+#endif
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+       bt_pages = FTL_Get_Block_Table_Flash_Size_Pages();
+
+       dwIPFPageAddr = g_wBlockTableOffset + bt_pages;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Writing IPF at "
+                              "Block %d Page %d\n",
+                              g_wBlockTableIndex, dwIPFPageAddr);
+
+#if CMD_DMA
+       wResult = GLOB_LLD_Write_Page_Main_Spare_cdma(g_pIPF,
+               g_wBlockTableIndex, dwIPFPageAddr, 1,
+               LLD_CMD_FLAG_MODE_CDMA | LLD_CMD_FLAG_ORDER_BEFORE_REST);
+       if (wResult == FAIL) {
+               nand_dbg_print(NAND_DBG_WARN,
+                              "NAND Program fail in %s, Line %d, "
+                              "Function: %s, new Bad Block %d generated!\n",
+                              __FILE__, __LINE__, __func__,
+                              g_wBlockTableIndex);
+       }
+       g_wBlockTableOffset = dwIPFPageAddr + 1;
+       p_BTableChangesDelta = (struct BTableChangesDelta *)g_pBTDelta_Free;
+       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+       p_BTableChangesDelta->ftl_cmd_cnt = ftl_cmd_cnt;
+       p_BTableChangesDelta->g_wBlockTableOffset = g_wBlockTableOffset;
+       p_BTableChangesDelta->ValidFields = 0x01;
+       ftl_cmd_cnt++;
+#else
+       wResult = GLOB_LLD_Write_Page_Main_Spare(g_pIPF,
+               g_wBlockTableIndex, dwIPFPageAddr, 1);
+       if (wResult == FAIL) {
+               nand_dbg_print(NAND_DBG_WARN,
+                              "NAND Program fail in %s, Line %d, "
+                              "Function: %s, new Bad Block %d generated!\n",
+                              __FILE__, __LINE__, __func__,
+                              (int)g_wBlockTableIndex);
+               MARK_BLOCK_AS_BAD(pbt[BLOCK_TABLE_INDEX]);
+               wTempBlockTableIndex = FTL_Replace_Block_Table();
+               bt_block_changed = 1;
+               if (BAD_BLOCK == wTempBlockTableIndex)
+                       return ERR;
+               g_wBlockTableIndex = wTempBlockTableIndex;
+               g_wBlockTableOffset = 0;
+               /* Block table tag is '00'. Means it's used one */
+               pbt[BLOCK_TABLE_INDEX] = g_wBlockTableIndex;
+               return FAIL;
+       }
+       g_wBlockTableOffset = dwIPFPageAddr + 1;
+#endif
+       return wResult;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     FTL_Read_Disturbance
+* Inputs:       block address
+* Outputs:      PASS=0 / FAIL=1
+* Description:  used to handle read disturbance. Data in block that
+*               reaches its read limit is moved to new block
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int FTL_Read_Disturbance(u32 blk_addr)
+{
+       int wResult = FAIL;
+       u32 *pbt = (u32 *) g_pBlockTable;
+       u32 dwOldBlockAddr = blk_addr;
+       u32 wBlockNum;
+       u32 i;
+       u32 wLeastReadCounter = 0xFFFF;
+       u32 wLeastReadIndex = BAD_BLOCK;
+       u32 wSpareBlockNum = 0;
+       u32 wTempNode;
+       u32 wReplacedNode;
+       u8 *g_pTempBuf;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                              __FILE__, __LINE__, __func__);
+
+#if CMD_DMA
+       g_pTempBuf = cp_back_buf_copies[cp_back_buf_idx];
+       cp_back_buf_idx++;
+       if (cp_back_buf_idx > COPY_BACK_BUF_NUM) {
+               printk(KERN_ERR "cp_back_buf_copies overflow! Exit."
+               "Maybe too many pending commands in your CDMA chain.\n");
+               return FAIL;
+       }
+#else
+       g_pTempBuf = tmp_buf_read_disturbance;
+#endif
+
+       wBlockNum = FTL_Get_Block_Index(blk_addr);
+
+       do {
+               /* This is a bug.Here 'i' should be logical block number
+                * and start from 1 (0 is reserved for block table).
+                * Have fixed it.        - Yunpeng 2008. 12. 19
+                */
+               for (i = 1; i < DeviceInfo.wDataBlockNum; i++) {
+                       if (IS_SPARE_BLOCK(i)) {
+                               u32 wPhysicalIndex =
+                                       (u32)((~SPARE_BLOCK) & pbt[i]);
+                               if (g_pReadCounter[wPhysicalIndex -
+                                       DeviceInfo.wSpectraStartBlock] <
+                                       wLeastReadCounter) {
+                                       wLeastReadCounter =
+                                               g_pReadCounter[wPhysicalIndex -
+                                               DeviceInfo.wSpectraStartBlock];
+                                       wLeastReadIndex = i;
+                               }
+                               wSpareBlockNum++;
+                       }
+               }
+
+               if (wSpareBlockNum <= NUM_FREE_BLOCKS_GATE) {
+                       wResult = GLOB_FTL_Garbage_Collection();
+                       if (PASS == wResult)
+                               continue;
+                       else
+                               break;
+               } else {
+                       wTempNode = (u32)(DISCARD_BLOCK | pbt[wBlockNum]);
+                       wReplacedNode = (u32)((~SPARE_BLOCK) &
+                                       pbt[wLeastReadIndex]);
+#if CMD_DMA
+                       pbt[wBlockNum] = wReplacedNode;
+                       pbt[wLeastReadIndex] = wTempNode;
+                       p_BTableChangesDelta =
+                               (struct BTableChangesDelta *)g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                       ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index = wBlockNum;
+                       p_BTableChangesDelta->BT_Entry_Value = pbt[wBlockNum];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+
+                       p_BTableChangesDelta =
+                               (struct BTableChangesDelta *)g_pBTDelta_Free;
+                       g_pBTDelta_Free += sizeof(struct BTableChangesDelta);
+
+                       p_BTableChangesDelta->ftl_cmd_cnt =
+                                       ftl_cmd_cnt;
+                       p_BTableChangesDelta->BT_Index = wLeastReadIndex;
+                       p_BTableChangesDelta->BT_Entry_Value =
+                                       pbt[wLeastReadIndex];
+                       p_BTableChangesDelta->ValidFields = 0x0C;
+
+                       wResult = GLOB_LLD_Read_Page_Main_cdma(g_pTempBuf,
+                               dwOldBlockAddr, 0, DeviceInfo.wPagesPerBlock,
+                               LLD_CMD_FLAG_MODE_CDMA);
+                       if (wResult == FAIL)
+                               return wResult;
+
+                       ftl_cmd_cnt++;
+
+                       if (wResult != FAIL) {
+                               if (FAIL == GLOB_LLD_Write_Page_Main_cdma(
+                                       g_pTempBuf, pbt[wBlockNum], 0,
+                                       DeviceInfo.wPagesPerBlock)) {
+                                       nand_dbg_print(NAND_DBG_WARN,
+                                               "NAND Program fail in "
+                                               "%s, Line %d, Function: %s, "
+                                               "new Bad Block %d "
+                                               "generated!\n",
+                                               __FILE__, __LINE__, __func__,
+                                               (int)pbt[wBlockNum]);
+                                       wResult = FAIL;
+                                       MARK_BLOCK_AS_BAD(pbt[wBlockNum]);
+                               }
+                               ftl_cmd_cnt++;
+                       }
+#else
+                       wResult = GLOB_LLD_Read_Page_Main(g_pTempBuf,
+                               dwOldBlockAddr, 0, DeviceInfo.wPagesPerBlock);
+                       if (wResult == FAIL)
+                               return wResult;
+
+                       if (wResult != FAIL) {
+                               /* This is a bug. At this time, pbt[wBlockNum]
+                               is still the physical address of
+                               discard block, and should not be write.
+                               Have fixed it as below.
+                                       -- Yunpeng 2008.12.19
+                               */
+                               wResult = GLOB_LLD_Write_Page_Main(g_pTempBuf,
+                                       wReplacedNode, 0,
+                                       DeviceInfo.wPagesPerBlock);
+                               if (wResult == FAIL) {
+                                       nand_dbg_print(NAND_DBG_WARN,
+                                               "NAND Program fail in "
+                                               "%s, Line %d, Function: %s, "
+                                               "new Bad Block %d "
+                                               "generated!\n",
+                                               __FILE__, __LINE__, __func__,
+                                               (int)wReplacedNode);
+                                       MARK_BLOCK_AS_BAD(wReplacedNode);
+                               } else {
+                                       pbt[wBlockNum] = wReplacedNode;
+                                       pbt[wLeastReadIndex] = wTempNode;
+                               }
+                       }
+
+                       if ((wResult == PASS) && (g_cBlockTableStatus !=
+                               IN_PROGRESS_BLOCK_TABLE)) {
+                               g_cBlockTableStatus = IN_PROGRESS_BLOCK_TABLE;
+                               FTL_Write_IN_Progress_Block_Table_Page();
+                       }
+#endif
+               }
+       } while (wResult != PASS)
+       ;
+
+#if CMD_DMA
+       /* ... */
+#endif
+
+       return wResult;
+}
+
diff --git a/drivers/block/spectra/flash.h b/drivers/block/spectra/flash.h
new file mode 100644 (file)
index 0000000..5ed0580
--- /dev/null
@@ -0,0 +1,198 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _FLASH_INTERFACE_
+#define _FLASH_INTERFACE_
+
+#include "ffsport.h"
+#include "spectraswconfig.h"
+
+#define MAX_BYTE_VALUE        0xFF
+#define MAX_WORD_VALUE        0xFFFF
+#define MAX_U32_VALUE        0xFFFFFFFF
+
+#define MAX_BLOCKNODE_VALUE     0xFFFFFF
+#define DISCARD_BLOCK           0x800000
+#define SPARE_BLOCK             0x400000
+#define BAD_BLOCK               0xC00000
+
+#define UNHIT_CACHE_ITEM         0xFFFF
+
+#define NAND_CACHE_INIT_ADDR    0xffffffffffffffffULL
+
+#define IN_PROGRESS_BLOCK_TABLE   0x00
+#define CURRENT_BLOCK_TABLE       0x01
+
+#define BTSIG_OFFSET   (0)
+#define BTSIG_BYTES    (5)
+#define BTSIG_DELTA    (3)
+
+#define MAX_READ_COUNTER  0x2710
+
+#define FIRST_BT_ID            (1)
+#define LAST_BT_ID    (254)
+#define BTBLOCK_INVAL  (u32)(0xFFFFFFFF)
+
+struct device_info_tag {
+       u16 wDeviceMaker;
+       u16 wDeviceID;
+       u32 wDeviceType;
+       u32 wSpectraStartBlock;
+       u32 wSpectraEndBlock;
+       u32 wTotalBlocks;
+       u16 wPagesPerBlock;
+       u16 wPageSize;
+       u16 wPageDataSize;
+       u16 wPageSpareSize;
+       u16 wNumPageSpareFlag;
+       u16 wECCBytesPerSector;
+       u32 wBlockSize;
+       u32 wBlockDataSize;
+       u32 wDataBlockNum;
+       u8 bPlaneNum;
+       u16 wDeviceMainAreaSize;
+       u16 wDeviceSpareAreaSize;
+       u16 wDevicesConnected;
+       u16 wDeviceWidth;
+       u16 wHWRevision;
+       u16 wHWFeatures;
+
+       u16 wONFIDevFeatures;
+       u16 wONFIOptCommands;
+       u16 wONFITimingMode;
+       u16 wONFIPgmCacheTimingMode;
+
+       u16 MLCDevice;
+       u16 wSpareSkipBytes;
+
+       u8 nBitsInPageNumber;
+       u8 nBitsInPageDataSize;
+       u8 nBitsInBlockDataSize;
+};
+
+extern struct device_info_tag DeviceInfo;
+
+/* Cache item format */
+struct flash_cache_item_tag {
+       u64 address;
+       u16 use_cnt;
+       u16 changed;
+       u8 *buf;
+};
+
+struct flash_cache_tag {
+       u32 cache_item_size; /* Size in bytes of each cache item */
+       u16 pages_per_item; /* How many NAND pages in each cache item */
+       u16 LRU; /* No. of the least recently used cache item */
+       struct flash_cache_item_tag array[CACHE_ITEM_NUM];
+};
+
+/*
+ *Data structure for each list node of the managment table
+ * used for the Level 2 Cache. Each node maps one logical NAND block.
+ */
+struct spectra_l2_cache_list {
+       struct list_head list;
+       u32 logical_blk_num; /* Logical block number */
+       u32 pages_array[]; /* Page map array of this logical block.
+                          * Array index is the logical block number,
+                          * and for every item of this arry:
+                          * high 16 bit is index of the L2 cache block num,
+                          * low 16 bit is the phy page num
+                          * of the above L2 cache block.
+                          * This array will be kmalloc during run time.
+                          */
+};
+
+struct spectra_l2_cache_info {
+       u32 blk_array[BLK_NUM_FOR_L2_CACHE];
+       u16 cur_blk_idx; /* idx to the phy block number of current using */
+       u16 cur_page_num; /* pages number of current using */
+       struct spectra_l2_cache_list table; /* First node of the table */
+};
+
+#define RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE    1
+
+#if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
+struct flash_cache_mod_item_tag {
+       u64 address;
+       u8 changed;
+};
+
+struct flash_cache_delta_list_tag {
+       u8 item;  /* used cache item */
+       struct flash_cache_mod_item_tag cache;
+};
+#endif
+
+extern struct flash_cache_tag Cache;
+
+extern u8 *buf_read_page_main_spare;
+extern u8 *buf_write_page_main_spare;
+extern u8 *buf_read_page_spare;
+extern u8 *buf_get_bad_block;
+extern u8 *cdma_desc_buf;
+extern u8 *memcp_desc_buf;
+
+/* struture used for IndentfyDevice function */
+struct spectra_indentfy_dev_tag {
+       u32 NumBlocks;
+       u16 PagesPerBlock;
+       u16 PageDataSize;
+       u16 wECCBytesPerSector;
+       u32 wDataBlockNum;
+};
+
+int GLOB_FTL_Flash_Init(void);
+int GLOB_FTL_Flash_Release(void);
+/*void GLOB_FTL_Erase_Flash(void);*/
+int GLOB_FTL_Block_Erase(u64 block_addr);
+int GLOB_FTL_Is_BadBlock(u32 block_num);
+int GLOB_FTL_IdentifyDevice(struct spectra_indentfy_dev_tag *dev_data);
+int GLOB_FTL_Event_Status(int *);
+u16 glob_ftl_execute_cmds(void);
+
+/*int FTL_Read_Disturbance(ADDRESSTYPE dwBlockAddr);*/
+int FTL_Read_Disturbance(u32 dwBlockAddr);
+
+/*Flash r/w based on cache*/
+int GLOB_FTL_Page_Read(u8 *read_data, u64 page_addr);
+int GLOB_FTL_Page_Write(u8 *write_data, u64 page_addr);
+int GLOB_FTL_Wear_Leveling(void);
+int GLOB_FTL_Flash_Format(void);
+int GLOB_FTL_Init(void);
+int GLOB_FTL_Flush_Cache(void);
+int GLOB_FTL_Garbage_Collection(void);
+int GLOB_FTL_BT_Garbage_Collection(void);
+void GLOB_FTL_Cache_Release(void);
+u8 *get_blk_table_start_addr(void);
+u8 *get_wear_leveling_table_start_addr(void);
+unsigned long get_blk_table_len(void);
+unsigned long get_wear_leveling_table_len(void);
+
+#if DEBUG_BNDRY
+void debug_boundary_lineno_error(int chnl, int limit, int no, int lineno,
+                               char *filename);
+#define debug_boundary_error(chnl, limit, no) debug_boundary_lineno_error(chnl,\
+                                               limit, no, __LINE__, __FILE__)
+#else
+#define debug_boundary_error(chnl, limit, no) ;
+#endif
+
+#endif /*_FLASH_INTERFACE_*/
diff --git a/drivers/block/spectra/lld.c b/drivers/block/spectra/lld.c
new file mode 100644 (file)
index 0000000..3f411af
--- /dev/null
@@ -0,0 +1,258 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include "spectraswconfig.h"
+#include "ffsport.h"
+#include "ffsdefs.h"
+#include "lld.h"
+#include "lld_nand.h"
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+#if FLASH_EMU          /* vector all the LLD calls to the LLD_EMU code */
+#include "lld_emu.h"
+#include "lld_cdma.h"
+
+/* common functions: */
+u16 GLOB_LLD_Flash_Reset(void)
+{
+       return emu_Flash_Reset();
+}
+
+u16 GLOB_LLD_Read_Device_ID(void)
+{
+       return emu_Read_Device_ID();
+}
+
+int GLOB_LLD_Flash_Release(void)
+{
+       return emu_Flash_Release();
+}
+
+u16 GLOB_LLD_Flash_Init(void)
+{
+       return emu_Flash_Init();
+}
+
+u16 GLOB_LLD_Erase_Block(u32 block_add)
+{
+       return emu_Erase_Block(block_add);
+}
+
+u16 GLOB_LLD_Write_Page_Main(u8 *write_data, u32 block, u16 Page,
+                               u16 PageCount)
+{
+       return emu_Write_Page_Main(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Main(u8 *read_data, u32 block, u16 Page,
+                              u16 PageCount)
+{
+       return emu_Read_Page_Main(read_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Main_Polling(u8 *read_data,
+                       u32 block, u16 page, u16 page_count)
+{
+       return emu_Read_Page_Main(read_data, block, page, page_count);
+}
+
+u16 GLOB_LLD_Write_Page_Main_Spare(u8 *write_data, u32 block,
+                                     u16 Page, u16 PageCount)
+{
+       return emu_Write_Page_Main_Spare(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Main_Spare(u8 *read_data, u32 block,
+                                    u16 Page, u16 PageCount)
+{
+       return emu_Read_Page_Main_Spare(read_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Write_Page_Spare(u8 *write_data, u32 block, u16 Page,
+                                u16 PageCount)
+{
+       return emu_Write_Page_Spare(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Spare(u8 *read_data, u32 block, u16 Page,
+                               u16 PageCount)
+{
+       return emu_Read_Page_Spare(read_data, block, Page, PageCount);
+}
+
+u16  GLOB_LLD_Get_Bad_Block(u32 block)
+{
+    return  emu_Get_Bad_Block(block);
+}
+
+#endif /* FLASH_EMU */
+
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+#if FLASH_NAND /* vector all the LLD calls to the NAND controller code */
+#include "lld_nand.h"
+#include "lld_cdma.h"
+#include "flash.h"
+
+/* common functions for LLD_NAND */
+void GLOB_LLD_ECC_Control(int enable)
+{
+       NAND_ECC_Ctrl(enable);
+}
+
+/* common functions for LLD_NAND */
+u16 GLOB_LLD_Flash_Reset(void)
+{
+       return NAND_Flash_Reset();
+}
+
+u16 GLOB_LLD_Read_Device_ID(void)
+{
+       return NAND_Read_Device_ID();
+}
+
+u16 GLOB_LLD_UnlockArrayAll(void)
+{
+       return NAND_UnlockArrayAll();
+}
+
+u16 GLOB_LLD_Flash_Init(void)
+{
+       return NAND_Flash_Init();
+}
+
+int GLOB_LLD_Flash_Release(void)
+{
+       return nand_release();
+}
+
+u16 GLOB_LLD_Erase_Block(u32 block_add)
+{
+       return NAND_Erase_Block(block_add);
+}
+
+
+u16 GLOB_LLD_Write_Page_Main(u8 *write_data, u32 block, u16 Page,
+                               u16 PageCount)
+{
+       return NAND_Write_Page_Main(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Main(u8 *read_data, u32 block, u16 page,
+                              u16 page_count)
+{
+       if (page_count == 1) /* Using polling to improve read speed */
+               return NAND_Read_Page_Main_Polling(read_data, block, page, 1);
+       else
+               return NAND_Read_Page_Main(read_data, block, page, page_count);
+}
+
+u16 GLOB_LLD_Read_Page_Main_Polling(u8 *read_data,
+                       u32 block, u16 page, u16 page_count)
+{
+       return NAND_Read_Page_Main_Polling(read_data,
+                       block, page, page_count);
+}
+
+u16 GLOB_LLD_Write_Page_Main_Spare(u8 *write_data, u32 block,
+                                     u16 Page, u16 PageCount)
+{
+       return NAND_Write_Page_Main_Spare(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Write_Page_Spare(u8 *write_data, u32 block, u16 Page,
+                                u16 PageCount)
+{
+       return NAND_Write_Page_Spare(write_data, block, Page, PageCount);
+}
+
+u16 GLOB_LLD_Read_Page_Main_Spare(u8 *read_data, u32 block,
+                                    u16 page, u16 page_count)
+{
+       return NAND_Read_Page_Main_Spare(read_data, block, page, page_count);
+}
+
+u16 GLOB_LLD_Read_Page_Spare(u8 *read_data, u32 block, u16 Page,
+                               u16 PageCount)
+{
+       return NAND_Read_Page_Spare(read_data, block, Page, PageCount);
+}
+
+u16  GLOB_LLD_Get_Bad_Block(u32 block)
+{
+       return  NAND_Get_Bad_Block(block);
+}
+
+u16 GLOB_LLD_Event_Status(void)
+{
+       return CDMA_Event_Status();
+}
+
+u16 glob_lld_execute_cmds(void)
+{
+       return CDMA_Execute_CMDs();
+}
+
+u16 GLOB_LLD_MemCopy_CMD(u8 *dest, u8 *src,
+                       u32 ByteCount, u16 flag)
+{
+       /* Replace the hardware memcopy with software memcpy function */
+       if (CDMA_Execute_CMDs())
+               return FAIL;
+       memcpy(dest, src, ByteCount);
+       return PASS;
+
+       /* return CDMA_MemCopy_CMD(dest, src, ByteCount, flag); */
+}
+
+u16 GLOB_LLD_Erase_Block_cdma(u32 block, u16 flags)
+{
+       return CDMA_Data_CMD(ERASE_CMD, 0, block, 0, 0, flags);
+}
+
+u16 GLOB_LLD_Write_Page_Main_cdma(u8 *data, u32 block, u16 page, u16 count)
+{
+       return CDMA_Data_CMD(WRITE_MAIN_CMD, data, block, page, count, 0);
+}
+
+u16 GLOB_LLD_Read_Page_Main_cdma(u8 *data, u32 block, u16 page,
+                               u16 count, u16 flags)
+{
+       return CDMA_Data_CMD(READ_MAIN_CMD, data, block, page, count, flags);
+}
+
+u16 GLOB_LLD_Write_Page_Main_Spare_cdma(u8 *data, u32 block, u16 page,
+                                       u16 count, u16 flags)
+{
+       return CDMA_Data_CMD(WRITE_MAIN_SPARE_CMD,
+                       data, block, page, count, flags);
+}
+
+u16 GLOB_LLD_Read_Page_Main_Spare_cdma(u8 *data,
+                               u32 block, u16 page, u16 count)
+{
+       return CDMA_Data_CMD(READ_MAIN_SPARE_CMD, data, block, page, count,
+                       LLD_CMD_FLAG_MODE_CDMA);
+}
+
+#endif /* FLASH_NAND */
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+
+/* end of LLD.c */
diff --git a/drivers/block/spectra/lld.h b/drivers/block/spectra/lld.h
new file mode 100644 (file)
index 0000000..d3738e0
--- /dev/null
@@ -0,0 +1,111 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+
+
+#ifndef _LLD_
+#define _LLD_
+
+#include "ffsport.h"
+#include "spectraswconfig.h"
+#include "flash.h"
+
+#define GOOD_BLOCK 0
+#define DEFECTIVE_BLOCK 1
+#define READ_ERROR 2
+
+#define CLK_X  5
+#define CLK_MULTI 4
+
+/* Typedefs */
+
+/*  prototypes: API for LLD */
+/* Currently, Write_Page_Main
+ *                       MemCopy
+ *                       Read_Page_Main_Spare
+ * do not have flag because they were not implemented prior to this
+ * They are not being added to keep changes to a minimum for now.
+ * Currently, they are not required (only reqd for Wr_P_M_S.)
+ * Later on, these NEED to be changed.
+ */
+
+extern void GLOB_LLD_ECC_Control(int enable);
+
+extern u16 GLOB_LLD_Flash_Reset(void);
+
+extern u16 GLOB_LLD_Read_Device_ID(void);
+
+extern u16 GLOB_LLD_UnlockArrayAll(void);
+
+extern u16 GLOB_LLD_Flash_Init(void);
+
+extern int GLOB_LLD_Flash_Release(void);
+
+extern u16 GLOB_LLD_Erase_Block(u32 block_add);
+
+extern u16 GLOB_LLD_Write_Page_Main(u8 *write_data,
+       u32 block, u16 Page, u16 PageCount);
+
+extern u16 GLOB_LLD_Read_Page_Main(u8 *read_data,
+       u32 block, u16 page, u16 page_count);
+
+extern u16 GLOB_LLD_Read_Page_Main_Polling(u8 *read_data,
+       u32 block, u16 page, u16 page_count);
+
+extern u16 GLOB_LLD_Write_Page_Main_Spare(u8 *write_data,
+       u32 block, u16 Page, u16 PageCount);
+
+extern u16 GLOB_LLD_Write_Page_Spare(u8 *write_data,
+       u32 block, u16 Page, u16 PageCount);
+
+extern u16 GLOB_LLD_Read_Page_Main_Spare(u8 *read_data,
+       u32 block, u16 page, u16 page_count);
+
+extern u16 GLOB_LLD_Read_Page_Spare(u8 *read_data,
+       u32 block, u16 Page, u16 PageCount);
+
+extern u16  GLOB_LLD_Get_Bad_Block(u32 block);
+
+extern u16 GLOB_LLD_Event_Status(void);
+
+extern u16 GLOB_LLD_MemCopy_CMD(u8 *dest, u8 *src, u32 ByteCount, u16 flag);
+
+extern u16 glob_lld_execute_cmds(void);
+
+extern u16 GLOB_LLD_Erase_Block_cdma(u32 block, u16 flags);
+
+extern u16 GLOB_LLD_Write_Page_Main_cdma(u8 *data,
+       u32 block, u16 page, u16 count);
+
+extern u16 GLOB_LLD_Read_Page_Main_cdma(u8 *data,
+       u32 block, u16 page, u16 count, u16 flags);
+
+extern u16 GLOB_LLD_Write_Page_Main_Spare_cdma(u8 *data,
+       u32 block, u16 page, u16 count, u16 flags);
+
+extern u16 GLOB_LLD_Read_Page_Main_Spare_cdma(u8 *data,
+       u32 block, u16 page, u16 count);
+
+#define LLD_CMD_FLAG_ORDER_BEFORE_REST         (0x1)
+#define LLD_CMD_FLAG_MODE_CDMA                 (0x8)
+
+
+#endif /*_LLD_ */
+
+
diff --git a/drivers/block/spectra/lld_cdma.c b/drivers/block/spectra/lld_cdma.c
new file mode 100644 (file)
index 0000000..c6e7610
--- /dev/null
@@ -0,0 +1,910 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include <linux/fs.h>
+#include <linux/slab.h>
+
+#include "spectraswconfig.h"
+#include "lld.h"
+#include "lld_nand.h"
+#include "lld_cdma.h"
+#include "lld_emu.h"
+#include "flash.h"
+#include "nand_regs.h"
+
+#define MAX_PENDING_CMDS    4
+#define MODE_02             (0x2 << 26)
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_Data_Cmd
+* Inputs:   cmd code (aligned for hw)
+*               data: pointer to source or destination
+*               block: block address
+*               page: page address
+*               num: num pages to transfer
+* Outputs:      PASS
+* Description:  This function takes the parameters and puts them
+*                   into the "pending commands" array.
+*               It does not parse or validate the parameters.
+*               The array index is same as the tag.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 CDMA_Data_CMD(u8 cmd, u8 *data, u32 block, u16 page, u16 num, u16 flags)
+{
+       u8 bank;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (0 == cmd)
+               nand_dbg_print(NAND_DBG_DEBUG,
+               "%s, Line %d, Illegal cmd (0)\n", __FILE__, __LINE__);
+
+       /* If a command of another bank comes, then first execute */
+       /* pending commands of the current bank, then set the new */
+       /* bank as current bank */
+       bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+       if (bank != info.flash_bank) {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Will access new bank. old bank: %d, new bank: %d\n",
+                       info.flash_bank, bank);
+               if (CDMA_Execute_CMDs()) {
+                       printk(KERN_ERR "CDMA_Execute_CMDs fail!\n");
+                       return FAIL;
+               }
+               info.flash_bank = bank;
+       }
+
+       info.pcmds[info.pcmds_num].CMD = cmd;
+       info.pcmds[info.pcmds_num].DataAddr = data;
+       info.pcmds[info.pcmds_num].Block = block;
+       info.pcmds[info.pcmds_num].Page = page;
+       info.pcmds[info.pcmds_num].PageCount = num;
+       info.pcmds[info.pcmds_num].DataDestAddr = 0;
+       info.pcmds[info.pcmds_num].DataSrcAddr = 0;
+       info.pcmds[info.pcmds_num].MemCopyByteCnt = 0;
+       info.pcmds[info.pcmds_num].Flags = flags;
+       info.pcmds[info.pcmds_num].Status = 0xB0B;
+
+       switch (cmd) {
+       case WRITE_MAIN_SPARE_CMD:
+               Conv_Main_Spare_Data_Log2Phy_Format(data, num);
+               break;
+       case WRITE_SPARE_CMD:
+               Conv_Spare_Data_Log2Phy_Format(data);
+               break;
+       default:
+               break;
+       }
+
+       info.pcmds_num++;
+
+       if (info.pcmds_num >= MAX_PENDING_CMDS) {
+               if (CDMA_Execute_CMDs()) {
+                       printk(KERN_ERR "CDMA_Execute_CMDs fail!\n");
+                       return FAIL;
+               }
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_MemCopy_CMD
+* Inputs:       dest: pointer to destination
+*               src:  pointer to source
+*               count: num bytes to transfer
+* Outputs:      PASS
+* Description:  This function takes the parameters and puts them
+*                   into the "pending commands" array.
+*               It does not parse or validate the parameters.
+*               The array index is same as the tag.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 CDMA_MemCopy_CMD(u8 *dest, u8 *src, u32 byte_cnt, u16 flags)
+{
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       info.pcmds[info.pcmds_num].CMD = MEMCOPY_CMD;
+       info.pcmds[info.pcmds_num].DataAddr = 0;
+       info.pcmds[info.pcmds_num].Block = 0;
+       info.pcmds[info.pcmds_num].Page = 0;
+       info.pcmds[info.pcmds_num].PageCount = 0;
+       info.pcmds[info.pcmds_num].DataDestAddr = dest;
+       info.pcmds[info.pcmds_num].DataSrcAddr = src;
+       info.pcmds[info.pcmds_num].MemCopyByteCnt = byte_cnt;
+       info.pcmds[info.pcmds_num].Flags = flags;
+       info.pcmds[info.pcmds_num].Status = 0xB0B;
+
+       info.pcmds_num++;
+
+       if (info.pcmds_num >= MAX_PENDING_CMDS) {
+               if (CDMA_Execute_CMDs()) {
+                       printk(KERN_ERR "CDMA_Execute_CMDs fail!\n");
+                       return FAIL;
+               }
+       }
+
+       return PASS;
+}
+
+#if 0
+/* Prints the PendingCMDs array */
+void print_pending_cmds(void)
+{
+       u16 i;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < info.pcmds_num; i++) {
+               nand_dbg_print(NAND_DBG_DEBUG, "\ni: %d\n", i);
+               switch (info.pcmds[i].CMD) {
+               case ERASE_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Erase Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case WRITE_MAIN_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Write Main Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case WRITE_MAIN_SPARE_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Write Main Spare Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case READ_MAIN_SPARE_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Read Main Spare Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case READ_MAIN_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Read Main Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case MEMCOPY_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Memcopy Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               case DUMMY_CMD:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Dummy Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               default:
+                       nand_dbg_print(NAND_DBG_DEBUG,
+                               "Illegal Command (0x%x)\n",
+                               info.pcmds[i].CMD);
+                       break;
+               }
+
+               nand_dbg_print(NAND_DBG_DEBUG, "DataAddr: 0x%x\n",
+                       (u32)info.pcmds[i].DataAddr);
+               nand_dbg_print(NAND_DBG_DEBUG, "Block: %d\n",
+                       info.pcmds[i].Block);
+               nand_dbg_print(NAND_DBG_DEBUG, "Page: %d\n",
+                       info.pcmds[i].Page);
+               nand_dbg_print(NAND_DBG_DEBUG, "PageCount: %d\n",
+                       info.pcmds[i].PageCount);
+               nand_dbg_print(NAND_DBG_DEBUG, "DataDestAddr: 0x%x\n",
+                       (u32)info.pcmds[i].DataDestAddr);
+               nand_dbg_print(NAND_DBG_DEBUG, "DataSrcAddr: 0x%x\n",
+                       (u32)info.pcmds[i].DataSrcAddr);
+               nand_dbg_print(NAND_DBG_DEBUG, "MemCopyByteCnt: %d\n",
+                       info.pcmds[i].MemCopyByteCnt);
+               nand_dbg_print(NAND_DBG_DEBUG, "Flags: 0x%x\n",
+                       info.pcmds[i].Flags);
+               nand_dbg_print(NAND_DBG_DEBUG, "Status: 0x%x\n",
+                       info.pcmds[i].Status);
+       }
+}
+
+/* Print the CDMA descriptors */
+void print_cdma_descriptors(void)
+{
+       struct cdma_descriptor *pc;
+       int i;
+
+       pc = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "\nWill dump cdma descriptors:\n");
+
+       for (i = 0; i < info.cdma_num; i++) {
+               nand_dbg_print(NAND_DBG_DEBUG, "\ni: %d\n", i);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "NxtPointerHi: 0x%x, NxtPointerLo: 0x%x\n",
+                       pc[i].NxtPointerHi, pc[i].NxtPointerLo);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "FlashPointerHi: 0x%x, FlashPointerLo: 0x%x\n",
+                       pc[i].FlashPointerHi, pc[i].FlashPointerLo);
+               nand_dbg_print(NAND_DBG_DEBUG, "CommandType: 0x%x\n",
+                       pc[i].CommandType);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "MemAddrHi: 0x%x, MemAddrLo: 0x%x\n",
+                       pc[i].MemAddrHi, pc[i].MemAddrLo);
+               nand_dbg_print(NAND_DBG_DEBUG, "CommandFlags: 0x%x\n",
+                       pc[i].CommandFlags);
+               nand_dbg_print(NAND_DBG_DEBUG, "Channel: %d, Status: 0x%x\n",
+                       pc[i].Channel, pc[i].Status);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "MemCopyPointerHi: 0x%x, MemCopyPointerLo: 0x%x\n",
+                       pc[i].MemCopyPointerHi, pc[i].MemCopyPointerLo);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Reserved12: 0x%x, Reserved13: 0x%x, "
+                       "Reserved14: 0x%x, pcmd: %d\n",
+                       pc[i].Reserved12, pc[i].Reserved13,
+                       pc[i].Reserved14, pc[i].pcmd);
+       }
+}
+
+/* Print the Memory copy descriptors */
+static void print_memcp_descriptors(void)
+{
+       struct memcpy_descriptor *pm;
+       int i;
+
+       pm = (struct memcpy_descriptor *)info.memcp_desc_buf;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "\nWill dump mem_cpy descriptors:\n");
+
+       for (i = 0; i < info.cdma_num; i++) {
+               nand_dbg_print(NAND_DBG_DEBUG, "\ni: %d\n", i);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "NxtPointerHi: 0x%x, NxtPointerLo: 0x%x\n",
+                       pm[i].NxtPointerHi, pm[i].NxtPointerLo);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "SrcAddrHi: 0x%x, SrcAddrLo: 0x%x\n",
+                       pm[i].SrcAddrHi, pm[i].SrcAddrLo);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "DestAddrHi: 0x%x, DestAddrLo: 0x%x\n",
+                       pm[i].DestAddrHi, pm[i].DestAddrLo);
+               nand_dbg_print(NAND_DBG_DEBUG, "XferSize: %d\n",
+                       pm[i].XferSize);
+               nand_dbg_print(NAND_DBG_DEBUG, "MemCopyFlags: 0x%x\n",
+                       pm[i].MemCopyFlags);
+               nand_dbg_print(NAND_DBG_DEBUG, "MemCopyStatus: %d\n",
+                       pm[i].MemCopyStatus);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved9: 0x%x\n",
+                       pm[i].reserved9);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved10: 0x%x\n",
+                       pm[i].reserved10);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved11: 0x%x\n",
+                       pm[i].reserved11);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved12: 0x%x\n",
+                       pm[i].reserved12);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved13: 0x%x\n",
+                       pm[i].reserved13);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved14: 0x%x\n",
+                       pm[i].reserved14);
+               nand_dbg_print(NAND_DBG_DEBUG, "reserved15: 0x%x\n",
+                       pm[i].reserved15);
+       }
+}
+#endif
+
+/* Reset cdma_descriptor chain to 0 */
+static void reset_cdma_desc(int i)
+{
+       struct cdma_descriptor *ptr;
+
+       BUG_ON(i >= MAX_DESCS);
+
+       ptr = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       ptr[i].NxtPointerHi = 0;
+       ptr[i].NxtPointerLo = 0;
+       ptr[i].FlashPointerHi = 0;
+       ptr[i].FlashPointerLo = 0;
+       ptr[i].CommandType = 0;
+       ptr[i].MemAddrHi = 0;
+       ptr[i].MemAddrLo = 0;
+       ptr[i].CommandFlags = 0;
+       ptr[i].Channel = 0;
+       ptr[i].Status = 0;
+       ptr[i].MemCopyPointerHi = 0;
+       ptr[i].MemCopyPointerLo = 0;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_UpdateEventStatus
+* Inputs:       none
+* Outputs:      none
+* Description:  This function update the event status of all the channels
+*               when an error condition is reported.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+void CDMA_UpdateEventStatus(void)
+{
+       int i, j, active_chan;
+       struct cdma_descriptor *ptr;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       ptr = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       for (j = 0; j < info.cdma_num; j++) {
+               /* Check for the descriptor with failure */
+               if ((ptr[j].Status & CMD_DMA_DESC_FAIL))
+                       break;
+
+       }
+
+       /* All the previous cmd's status for this channel must be good */
+       for (i = 0; i < j; i++) {
+               if (ptr[i].pcmd != 0xff)
+                       info.pcmds[ptr[i].pcmd].Status = CMD_PASS;
+       }
+
+       /* Abort the channel with type 0 reset command. It resets the */
+       /* selected channel after the descriptor completes the flash */
+       /* operation and status has been updated for the descriptor. */
+       /* Memory Copy and Sync associated with this descriptor will */
+       /* not be executed */
+       active_chan = ioread32(FlashReg + CHNL_ACTIVE);
+       if ((active_chan & (1 << info.flash_bank)) == (1 << info.flash_bank)) {
+               iowrite32(MODE_02 | (0 << 4), FlashMem); /* Type 0 reset */
+               iowrite32((0xF << 4) | info.flash_bank, FlashMem + 0x10);
+       } else { /* Should not reached here */
+               printk(KERN_ERR "Error! Used bank is not set in"
+                       " reg CHNL_ACTIVE\n");
+       }
+}
+
+static void cdma_trans(u16 chan)
+{
+       u32 addr;
+
+       addr = info.cdma_desc;
+
+       iowrite32(MODE_10 | (chan << 24), FlashMem);
+       iowrite32((1 << 7) | chan, FlashMem + 0x10);
+
+       iowrite32(MODE_10 | (chan << 24) | ((0x0FFFF & (addr >> 16)) << 8),
+               FlashMem);
+       iowrite32((1 << 7) | (1 << 4) | 0, FlashMem + 0x10);
+
+       iowrite32(MODE_10 | (chan << 24) | ((0x0FFFF & addr) << 8), FlashMem);
+       iowrite32((1 << 7) | (1 << 5) | 0, FlashMem + 0x10);
+
+       iowrite32(MODE_10 | (chan << 24), FlashMem);
+       iowrite32((1 << 7) | (1 << 5) | (1 << 4) | 0, FlashMem + 0x10);
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_Execute_CMDs (for use with CMD_DMA)
+* Inputs:       tag_count:  the number of pending cmds to do
+* Outputs:      PASS/FAIL
+* Description:  Build the SDMA chain(s) by making one CMD-DMA descriptor
+*               for each pending command, start the CDMA engine, and return.
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 CDMA_Execute_CMDs(void)
+{
+       int i, ret;
+       u64 flash_add;
+       u32 ptr;
+       dma_addr_t map_addr, next_ptr;
+       u16 status = PASS;
+       u16 tmp_c;
+       struct cdma_descriptor *pc;
+       struct memcpy_descriptor *pm;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       /* No pending cmds to execute, just exit */
+       if (0 == info.pcmds_num) {
+               nand_dbg_print(NAND_DBG_TRACE,
+                       "No pending cmds to execute. Just exit.\n");
+               return PASS;
+       }
+
+       for (i = 0; i < MAX_DESCS; i++)
+               reset_cdma_desc(i);
+
+       pc = (struct cdma_descriptor *)info.cdma_desc_buf;
+       pm = (struct memcpy_descriptor *)info.memcp_desc_buf;
+
+       info.cdma_desc = virt_to_bus(info.cdma_desc_buf);
+       info.memcp_desc = virt_to_bus(info.memcp_desc_buf);
+       next_ptr = info.cdma_desc;
+       info.cdma_num = 0;
+
+       for (i = 0; i < info.pcmds_num; i++) {
+               if (info.pcmds[i].Block >= DeviceInfo.wTotalBlocks) {
+                       info.pcmds[i].Status = CMD_NOT_DONE;
+                       continue;
+               }
+
+               next_ptr += sizeof(struct cdma_descriptor);
+               pc[info.cdma_num].NxtPointerHi = next_ptr >> 16;
+               pc[info.cdma_num].NxtPointerLo = next_ptr & 0xffff;
+
+               /* Use the Block offset within a bank */
+               tmp_c = info.pcmds[i].Block /
+                       (DeviceInfo.wTotalBlocks / totalUsedBanks);
+               flash_add = (u64)(info.pcmds[i].Block - tmp_c *
+                       (DeviceInfo.wTotalBlocks / totalUsedBanks)) *
+                       DeviceInfo.wBlockDataSize +
+                       (u64)(info.pcmds[i].Page) *
+                       DeviceInfo.wPageDataSize;
+
+               ptr = MODE_10 | (info.flash_bank << 24) |
+                       (u32)GLOB_u64_Div(flash_add,
+                               DeviceInfo.wPageDataSize);
+               pc[info.cdma_num].FlashPointerHi = ptr >> 16;
+               pc[info.cdma_num].FlashPointerLo = ptr & 0xffff;
+
+               if ((info.pcmds[i].CMD == WRITE_MAIN_SPARE_CMD) ||
+                       (info.pcmds[i].CMD == READ_MAIN_SPARE_CMD)) {
+                       /* Descriptor to set Main+Spare Access Mode */
+                       pc[info.cdma_num].CommandType = 0x43;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       pc[info.cdma_num].MemAddrHi = 0;
+                       pc[info.cdma_num].MemAddrLo = 0;
+                       pc[info.cdma_num].Channel = 0;
+                       pc[info.cdma_num].Status = 0;
+                       pc[info.cdma_num].pcmd = i;
+
+                       info.cdma_num++;
+                       BUG_ON(info.cdma_num >= MAX_DESCS);
+
+                       reset_cdma_desc(info.cdma_num);
+                       next_ptr += sizeof(struct cdma_descriptor);
+                       pc[info.cdma_num].NxtPointerHi = next_ptr >> 16;
+                       pc[info.cdma_num].NxtPointerLo = next_ptr & 0xffff;
+                       pc[info.cdma_num].FlashPointerHi = ptr >> 16;
+                       pc[info.cdma_num].FlashPointerLo = ptr & 0xffff;
+               }
+
+               switch (info.pcmds[i].CMD) {
+               case ERASE_CMD:
+                       pc[info.cdma_num].CommandType = 1;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       pc[info.cdma_num].MemAddrHi = 0;
+                       pc[info.cdma_num].MemAddrLo = 0;
+                       break;
+
+               case WRITE_MAIN_CMD:
+                       pc[info.cdma_num].CommandType =
+                               0x2100 | info.pcmds[i].PageCount;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       map_addr = virt_to_bus(info.pcmds[i].DataAddr);
+                       pc[info.cdma_num].MemAddrHi = map_addr >> 16;
+                       pc[info.cdma_num].MemAddrLo = map_addr & 0xffff;
+                       break;
+
+               case READ_MAIN_CMD:
+                       pc[info.cdma_num].CommandType =
+                               0x2000 | info.pcmds[i].PageCount;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       map_addr = virt_to_bus(info.pcmds[i].DataAddr);
+                       pc[info.cdma_num].MemAddrHi = map_addr >> 16;
+                       pc[info.cdma_num].MemAddrLo = map_addr & 0xffff;
+                       break;
+
+               case WRITE_MAIN_SPARE_CMD:
+                       pc[info.cdma_num].CommandType =
+                               0x2100 | info.pcmds[i].PageCount;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       map_addr = virt_to_bus(info.pcmds[i].DataAddr);
+                       pc[info.cdma_num].MemAddrHi = map_addr >> 16;
+                       pc[info.cdma_num].MemAddrLo = map_addr & 0xffff;
+                       break;
+
+               case READ_MAIN_SPARE_CMD:
+                       pc[info.cdma_num].CommandType =
+                               0x2000 | info.pcmds[i].PageCount;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       map_addr = virt_to_bus(info.pcmds[i].DataAddr);
+                       pc[info.cdma_num].MemAddrHi = map_addr >> 16;
+                       pc[info.cdma_num].MemAddrLo = map_addr & 0xffff;
+                       break;
+
+               case MEMCOPY_CMD:
+                       pc[info.cdma_num].CommandType = 0xFFFF; /* NOP cmd */
+                       /* Set bit 11 to let the CDMA engine continue to */
+                       /* execute only after it has finished processing   */
+                       /* the memcopy descriptor.                        */
+                       /* Also set bit 10 and bit 9 to 1                  */
+                       pc[info.cdma_num].CommandFlags = 0x0E40;
+                       map_addr = info.memcp_desc + info.cdma_num *
+                                       sizeof(struct memcpy_descriptor);
+                       pc[info.cdma_num].MemCopyPointerHi = map_addr >> 16;
+                       pc[info.cdma_num].MemCopyPointerLo = map_addr & 0xffff;
+
+                       pm[info.cdma_num].NxtPointerHi = 0;
+                       pm[info.cdma_num].NxtPointerLo = 0;
+
+                       map_addr = virt_to_bus(info.pcmds[i].DataSrcAddr);
+                       pm[info.cdma_num].SrcAddrHi = map_addr >> 16;
+                       pm[info.cdma_num].SrcAddrLo = map_addr & 0xffff;
+                       map_addr = virt_to_bus(info.pcmds[i].DataDestAddr);
+                       pm[info.cdma_num].DestAddrHi = map_addr >> 16;
+                       pm[info.cdma_num].DestAddrLo = map_addr & 0xffff;
+
+                       pm[info.cdma_num].XferSize =
+                               info.pcmds[i].MemCopyByteCnt;
+                       pm[info.cdma_num].MemCopyFlags =
+                               (0 << 15 | 0 << 14 | 27 << 8 | 0x40);
+                       pm[info.cdma_num].MemCopyStatus = 0;
+                       break;
+
+               case DUMMY_CMD:
+               default:
+                       pc[info.cdma_num].CommandType = 0XFFFF;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       pc[info.cdma_num].MemAddrHi = 0;
+                       pc[info.cdma_num].MemAddrLo = 0;
+                       break;
+               }
+
+               pc[info.cdma_num].Channel = 0;
+               pc[info.cdma_num].Status = 0;
+               pc[info.cdma_num].pcmd = i;
+
+               info.cdma_num++;
+               BUG_ON(info.cdma_num >= MAX_DESCS);
+
+               if ((info.pcmds[i].CMD == WRITE_MAIN_SPARE_CMD) ||
+                       (info.pcmds[i].CMD == READ_MAIN_SPARE_CMD)) {
+                       /* Descriptor to set back Main Area Access Mode */
+                       reset_cdma_desc(info.cdma_num);
+                       next_ptr += sizeof(struct cdma_descriptor);
+                       pc[info.cdma_num].NxtPointerHi = next_ptr >> 16;
+                       pc[info.cdma_num].NxtPointerLo = next_ptr & 0xffff;
+
+                       pc[info.cdma_num].FlashPointerHi = ptr >> 16;
+                       pc[info.cdma_num].FlashPointerLo = ptr & 0xffff;
+
+                       pc[info.cdma_num].CommandType = 0x42;
+                       pc[info.cdma_num].CommandFlags =
+                               (0 << 10) | (1 << 9) | (0 << 8) | 0x40;
+                       pc[info.cdma_num].MemAddrHi = 0;
+                       pc[info.cdma_num].MemAddrLo = 0;
+
+                       pc[info.cdma_num].Channel = 0;
+                       pc[info.cdma_num].Status = 0;
+                       pc[info.cdma_num].pcmd = i;
+
+                       info.cdma_num++;
+                       BUG_ON(info.cdma_num >= MAX_DESCS);
+               }
+       }
+
+       /* Add a dummy descriptor at end of the CDMA chain */
+       reset_cdma_desc(info.cdma_num);
+       ptr = MODE_10 | (info.flash_bank << 24);
+       pc[info.cdma_num].FlashPointerHi = ptr >> 16;
+       pc[info.cdma_num].FlashPointerLo = ptr & 0xffff;
+       pc[info.cdma_num].CommandType = 0xFFFF; /* NOP command */
+       /* Set Command Flags for the last CDMA descriptor: */
+       /* set Continue bit (bit 9) to 0 and Interrupt bit (bit 8) to 1 */
+       pc[info.cdma_num].CommandFlags =
+               (0 << 10) | (0 << 9) | (1 << 8) | 0x40;
+       pc[info.cdma_num].pcmd = 0xff; /* Set it to an illegal value */
+       info.cdma_num++;
+       BUG_ON(info.cdma_num >= MAX_DESCS);
+
+       iowrite32(1, FlashReg + GLOBAL_INT_ENABLE);  /* Enable Interrupt */
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       /* Wait for DMA to be enabled before issuing the next command */
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+       cdma_trans(info.flash_bank);
+
+       ret = wait_for_completion_timeout(&info.complete, 50 * HZ);
+       if (!ret)
+               printk(KERN_ERR "Wait for completion timeout "
+                       "in %s, Line %d\n", __FILE__, __LINE__);
+       status = info.ret;
+
+       info.pcmds_num = 0; /* Clear the pending cmds number to 0 */
+
+       return status;
+}
+
+int is_cdma_interrupt(void)
+{
+       u32 ints_b0, ints_b1, ints_b2, ints_b3, ints_cdma;
+       u32 int_en_mask;
+       u32 cdma_int_en_mask;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       /* Set the global Enable masks for only those interrupts
+        * that are supported */
+       cdma_int_en_mask = (DMA_INTR__DESC_COMP_CHANNEL0 |
+                       DMA_INTR__DESC_COMP_CHANNEL1 |
+                       DMA_INTR__DESC_COMP_CHANNEL2 |
+                       DMA_INTR__DESC_COMP_CHANNEL3 |
+                       DMA_INTR__MEMCOPY_DESC_COMP);
+
+       int_en_mask = (INTR_STATUS0__ECC_ERR |
+               INTR_STATUS0__PROGRAM_FAIL |
+               INTR_STATUS0__ERASE_FAIL);
+
+       ints_b0 = ioread32(FlashReg + INTR_STATUS0) & int_en_mask;
+       ints_b1 = ioread32(FlashReg + INTR_STATUS1) & int_en_mask;
+       ints_b2 = ioread32(FlashReg + INTR_STATUS2) & int_en_mask;
+       ints_b3 = ioread32(FlashReg + INTR_STATUS3) & int_en_mask;
+       ints_cdma = ioread32(FlashReg + DMA_INTR) & cdma_int_en_mask;
+
+       nand_dbg_print(NAND_DBG_WARN, "ints_bank0 to ints_bank3: "
+                       "0x%x, 0x%x, 0x%x, 0x%x, ints_cdma: 0x%x\n",
+                       ints_b0, ints_b1, ints_b2, ints_b3, ints_cdma);
+
+       if (ints_b0 || ints_b1 || ints_b2 || ints_b3 || ints_cdma) {
+               return 1;
+       } else {
+               iowrite32(ints_b0, FlashReg + INTR_STATUS0);
+               iowrite32(ints_b1, FlashReg + INTR_STATUS1);
+               iowrite32(ints_b2, FlashReg + INTR_STATUS2);
+               iowrite32(ints_b3, FlashReg + INTR_STATUS3);
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Not a NAND controller interrupt! Ignore it.\n");
+               return 0;
+       }
+}
+
+static void update_event_status(void)
+{
+       int i;
+       struct cdma_descriptor *ptr;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       ptr = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       for (i = 0; i < info.cdma_num; i++) {
+               if (ptr[i].pcmd != 0xff)
+                       info.pcmds[ptr[i].pcmd].Status = CMD_PASS;
+               if ((ptr[i].CommandType == 0x41) ||
+                       (ptr[i].CommandType == 0x42) ||
+                       (ptr[i].CommandType == 0x43))
+                       continue;
+
+               switch (info.pcmds[ptr[i].pcmd].CMD) {
+               case READ_MAIN_SPARE_CMD:
+                       Conv_Main_Spare_Data_Phy2Log_Format(
+                               info.pcmds[ptr[i].pcmd].DataAddr,
+                               info.pcmds[ptr[i].pcmd].PageCount);
+                       break;
+               case READ_SPARE_CMD:
+                       Conv_Spare_Data_Phy2Log_Format(
+                               info.pcmds[ptr[i].pcmd].DataAddr);
+                       break;
+               }
+       }
+}
+
+static u16 do_ecc_for_desc(u32 ch, u8 *buf, u16 page)
+{
+       u16 event = EVENT_NONE;
+       u16 err_byte;
+       u16 err_page = 0;
+       u8 err_sector;
+       u8 err_device;
+       u16 ecc_correction_info;
+       u16 err_address;
+       u32 eccSectorSize;
+       u8 *err_pos;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       eccSectorSize = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+
+       do {
+               if (0 == ch)
+                       err_page = ioread32(FlashReg + ERR_PAGE_ADDR0);
+               else if (1 == ch)
+                       err_page = ioread32(FlashReg + ERR_PAGE_ADDR1);
+               else if (2 == ch)
+                       err_page = ioread32(FlashReg + ERR_PAGE_ADDR2);
+               else if (3 == ch)
+                       err_page = ioread32(FlashReg + ERR_PAGE_ADDR3);
+
+               err_address = ioread32(FlashReg + ECC_ERROR_ADDRESS);
+               err_byte = err_address & ECC_ERROR_ADDRESS__OFFSET;
+               err_sector = ((err_address &
+                       ECC_ERROR_ADDRESS__SECTOR_NR) >> 12);
+
+               ecc_correction_info = ioread32(FlashReg + ERR_CORRECTION_INFO);
+               err_device = ((ecc_correction_info &
+                       ERR_CORRECTION_INFO__DEVICE_NR) >> 8);
+
+               if (ecc_correction_info & ERR_CORRECTION_INFO__ERROR_TYPE) {
+                       event = EVENT_UNCORRECTABLE_DATA_ERROR;
+               } else {
+                       event = EVENT_CORRECTABLE_DATA_ERROR_FIXED;
+                       if (err_byte < ECC_SECTOR_SIZE) {
+                               err_pos = buf +
+                                       (err_page - page) *
+                                       DeviceInfo.wPageDataSize +
+                                       err_sector * eccSectorSize +
+                                       err_byte *
+                                       DeviceInfo.wDevicesConnected +
+                                       err_device;
+                               *err_pos ^= ecc_correction_info &
+                                       ERR_CORRECTION_INFO__BYTEMASK;
+                       }
+               }
+       } while (!(ecc_correction_info & ERR_CORRECTION_INFO__LAST_ERR_INFO));
+
+       return event;
+}
+
+static u16 process_ecc_int(u32 c, u16 *p_desc_num)
+{
+       struct cdma_descriptor *ptr;
+       u16 j;
+       int event = EVENT_PASS;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (c != info.flash_bank)
+               printk(KERN_ERR "Error!info.flash_bank is %d, while c is %d\n",
+                               info.flash_bank, c);
+
+       ptr = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       for (j = 0; j < info.cdma_num; j++)
+               if ((ptr[j].Status & CMD_DMA_DESC_COMP) != CMD_DMA_DESC_COMP)
+                       break;
+
+       *p_desc_num = j; /* Pass the descripter number found here */
+
+       if (j >= info.cdma_num) {
+               printk(KERN_ERR "Can not find the correct descriptor number "
+                       "when ecc interrupt triggered!"
+                       "info.cdma_num: %d, j: %d\n", info.cdma_num, j);
+               return EVENT_UNCORRECTABLE_DATA_ERROR;
+       }
+
+       event = do_ecc_for_desc(c, info.pcmds[ptr[j].pcmd].DataAddr,
+               info.pcmds[ptr[j].pcmd].Page);
+
+       if (EVENT_UNCORRECTABLE_DATA_ERROR == event) {
+               printk(KERN_ERR "Uncorrectable ECC error!"
+                       "info.cdma_num: %d, j: %d, "
+                       "pending cmd CMD: 0x%x, "
+                       "Block: 0x%x, Page: 0x%x, PageCount: 0x%x\n",
+                       info.cdma_num, j,
+                       info.pcmds[ptr[j].pcmd].CMD,
+                       info.pcmds[ptr[j].pcmd].Block,
+                       info.pcmds[ptr[j].pcmd].Page,
+                       info.pcmds[ptr[j].pcmd].PageCount);
+
+               if (ptr[j].pcmd != 0xff)
+                       info.pcmds[ptr[j].pcmd].Status = CMD_FAIL;
+               CDMA_UpdateEventStatus();
+       }
+
+       return event;
+}
+
+static void process_prog_erase_fail_int(u16 desc_num)
+{
+       struct cdma_descriptor *ptr;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       ptr = (struct cdma_descriptor *)info.cdma_desc_buf;
+
+       if (ptr[desc_num].pcmd != 0xFF)
+               info.pcmds[ptr[desc_num].pcmd].Status = CMD_FAIL;
+
+       CDMA_UpdateEventStatus();
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_Event_Status (for use with CMD_DMA)
+* Inputs:       none
+* Outputs:      Event_Status code
+* Description:  This function is called after an interrupt has happened
+*               It reads the HW status register and ...tbd
+*               It returns the appropriate event status
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16  CDMA_Event_Status(void)
+{
+       u32 ints_addr[4] = {INTR_STATUS0, INTR_STATUS1,
+               INTR_STATUS2, INTR_STATUS3};
+       u32 dma_intr_bit[4] = {DMA_INTR__DESC_COMP_CHANNEL0,
+               DMA_INTR__DESC_COMP_CHANNEL1,
+               DMA_INTR__DESC_COMP_CHANNEL2,
+               DMA_INTR__DESC_COMP_CHANNEL3};
+       u32 cdma_int_status, int_status;
+       u32 ecc_enable = 0;
+       u16 event = EVENT_PASS;
+       u16 cur_desc = 0;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       ecc_enable = ioread32(FlashReg + ECC_ENABLE);
+
+       while (1) {
+               int_status = ioread32(FlashReg + ints_addr[info.flash_bank]);
+               if (ecc_enable && (int_status & INTR_STATUS0__ECC_ERR)) {
+                       event = process_ecc_int(info.flash_bank, &cur_desc);
+                       iowrite32(INTR_STATUS0__ECC_ERR,
+                               FlashReg + ints_addr[info.flash_bank]);
+                       if (EVENT_UNCORRECTABLE_DATA_ERROR == event) {
+                               nand_dbg_print(NAND_DBG_WARN,
+                                       "ints_bank0 to ints_bank3: "
+                                       "0x%x, 0x%x, 0x%x, 0x%x, "
+                                       "ints_cdma: 0x%x\n",
+                                       ioread32(FlashReg + INTR_STATUS0),
+                                       ioread32(FlashReg + INTR_STATUS1),
+                                       ioread32(FlashReg + INTR_STATUS2),
+                                       ioread32(FlashReg + INTR_STATUS3),
+                                       ioread32(FlashReg + DMA_INTR));
+                               break;
+                       }
+               } else if (int_status & INTR_STATUS0__PROGRAM_FAIL) {
+                       printk(KERN_ERR "NAND program fail interrupt!\n");
+                       process_prog_erase_fail_int(cur_desc);
+                       event = EVENT_PROGRAM_FAILURE;
+                       break;
+               } else if (int_status & INTR_STATUS0__ERASE_FAIL) {
+                       printk(KERN_ERR "NAND erase fail interrupt!\n");
+                       process_prog_erase_fail_int(cur_desc);
+                       event = EVENT_ERASE_FAILURE;
+                       break;
+               } else {
+                       cdma_int_status = ioread32(FlashReg + DMA_INTR);
+                       if (cdma_int_status & dma_intr_bit[info.flash_bank]) {
+                               iowrite32(dma_intr_bit[info.flash_bank],
+                                       FlashReg + DMA_INTR);
+                               update_event_status();
+                               event = EVENT_PASS;
+                               break;
+                       }
+               }
+       }
+
+       int_status = ioread32(FlashReg + ints_addr[info.flash_bank]);
+       iowrite32(int_status, FlashReg + ints_addr[info.flash_bank]);
+       cdma_int_status = ioread32(FlashReg + DMA_INTR);
+       iowrite32(cdma_int_status, FlashReg + DMA_INTR);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       return event;
+}
+
+
+
diff --git a/drivers/block/spectra/lld_cdma.h b/drivers/block/spectra/lld_cdma.h
new file mode 100644 (file)
index 0000000..854ea06
--- /dev/null
@@ -0,0 +1,123 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+/* header for LLD_CDMA.c module */
+
+#ifndef _LLD_CDMA_
+#define _LLD_CDMA_
+
+#include "flash.h"
+
+#define  DEBUG_SYNC    1
+
+/*///////////   CDMA specific MACRO definition */
+#define MAX_DESCS         (255)
+#define MAX_CHANS  (4)
+#define MAX_SYNC_POINTS         (16)
+#define MAX_DESC_PER_CHAN     (MAX_DESCS * 3 + MAX_SYNC_POINTS + 2)
+
+#define CHANNEL_SYNC_MASK       (0x000F)
+#define CHANNEL_DMA_MASK        (0x00F0)
+#define CHANNEL_ID_MASK         (0x0300)
+#define CHANNEL_CONT_MASK       (0x4000)
+#define CHANNEL_INTR_MASK       (0x8000)
+
+#define CHANNEL_SYNC_OFFSET     (0)
+#define CHANNEL_DMA_OFFSET      (4)
+#define CHANNEL_ID_OFFSET       (8)
+#define CHANNEL_CONT_OFFSET     (14)
+#define CHANNEL_INTR_OFFSET     (15)
+
+u16 CDMA_Data_CMD(u8 cmd, u8 *data, u32 block, u16 page, u16 num, u16 flags);
+u16 CDMA_MemCopy_CMD(u8 *dest, u8 *src, u32 byte_cnt, u16 flags);
+u16 CDMA_Execute_CMDs(void);
+void print_pending_cmds(void);
+void print_cdma_descriptors(void);
+
+extern u8 g_SBDCmdIndex;
+extern struct mrst_nand_info info;
+
+
+/*///////////   prototypes: APIs for LLD_CDMA */
+int is_cdma_interrupt(void);
+u16 CDMA_Event_Status(void);
+
+/* CMD-DMA Descriptor Struct.  These are defined by the CMD_DMA HW */
+struct cdma_descriptor {
+       u32 NxtPointerHi;
+       u32 NxtPointerLo;
+       u32 FlashPointerHi;
+       u32 FlashPointerLo;
+       u32 CommandType;
+       u32 MemAddrHi;
+       u32 MemAddrLo;
+       u32 CommandFlags;
+       u32 Channel;
+       u32 Status;
+       u32 MemCopyPointerHi;
+       u32 MemCopyPointerLo;
+       u32 Reserved12;
+       u32 Reserved13;
+       u32 Reserved14;
+       u32 pcmd; /* pending cmd num related to this descriptor */
+};
+
+/* This struct holds one MemCopy descriptor as defined by the HW */
+struct memcpy_descriptor {
+       u32 NxtPointerHi;
+       u32 NxtPointerLo;
+       u32 SrcAddrHi;
+       u32 SrcAddrLo;
+       u32 DestAddrHi;
+       u32 DestAddrLo;
+       u32 XferSize;
+       u32 MemCopyFlags;
+       u32 MemCopyStatus;
+       u32 reserved9;
+       u32 reserved10;
+       u32 reserved11;
+       u32 reserved12;
+       u32 reserved13;
+       u32 reserved14;
+       u32 reserved15;
+};
+
+/* Pending CMD table entries (includes MemCopy parameters */
+struct pending_cmd {
+       u8 CMD;
+       u8 *DataAddr;
+       u32 Block;
+       u16 Page;
+       u16 PageCount;
+       u8 *DataDestAddr;
+       u8 *DataSrcAddr;
+       u32 MemCopyByteCnt;
+       u16 Flags;
+       u16 Status;
+};
+
+#if DEBUG_SYNC
+extern u32 debug_sync_cnt;
+#endif
+
+/* Definitions for CMD DMA descriptor chain fields */
+#define     CMD_DMA_DESC_COMP   0x8000
+#define     CMD_DMA_DESC_FAIL   0x4000
+
+#endif /*_LLD_CDMA_*/
diff --git a/drivers/block/spectra/lld_emu.c b/drivers/block/spectra/lld_emu.c
new file mode 100644 (file)
index 0000000..60eb0f6
--- /dev/null
@@ -0,0 +1,780 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include <linux/fs.h>
+#include <linux/slab.h>
+#include "flash.h"
+#include "ffsdefs.h"
+#include "lld_emu.h"
+#include "lld.h"
+#if CMD_DMA
+#include "lld_cdma.h"
+#endif
+
+#define GLOB_LLD_PAGES           64
+#define GLOB_LLD_PAGE_SIZE       (512+16)
+#define GLOB_LLD_PAGE_DATA_SIZE  512
+#define GLOB_LLD_BLOCKS          2048
+
+#if (CMD_DMA  && FLASH_EMU)
+#include "lld_cdma.h"
+u32 totalUsedBanks;
+u32 valid_banks[MAX_CHANS];
+#endif
+
+#if FLASH_EMU                  /* This is for entire module */
+
+static u8 *flash_memory[GLOB_LLD_BLOCKS * GLOB_LLD_PAGES];
+
+/* Read nand emu file and then fill it's content to flash_memory */
+int emu_load_file_to_mem(void)
+{
+       mm_segment_t fs;
+       struct file *nef_filp = NULL;
+       struct inode *inode = NULL;
+       loff_t nef_size = 0;
+       loff_t tmp_file_offset, file_offset;
+       ssize_t nread;
+       int i, rc = -EINVAL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       fs = get_fs();
+       set_fs(get_ds());
+
+       nef_filp = filp_open("/root/nand_emu_file", O_RDWR | O_LARGEFILE, 0);
+       if (IS_ERR(nef_filp)) {
+               printk(KERN_ERR "filp_open error: "
+                      "Unable to open nand emu file!\n");
+               return PTR_ERR(nef_filp);
+       }
+
+       if (nef_filp->f_path.dentry) {
+               inode = nef_filp->f_path.dentry->d_inode;
+       } else {
+               printk(KERN_ERR "Can not get valid inode!\n");
+               goto out;
+       }
+
+       nef_size = i_size_read(inode->i_mapping->host);
+       if (nef_size <= 0) {
+               printk(KERN_ERR "Invalid nand emu file size: "
+                      "0x%llx\n", nef_size);
+               goto out;
+       } else {
+               nand_dbg_print(NAND_DBG_DEBUG, "nand emu file size: %lld\n",
+                              nef_size);
+       }
+
+       file_offset = 0;
+       for (i = 0; i < GLOB_LLD_BLOCKS * GLOB_LLD_PAGES; i++) {
+               tmp_file_offset = file_offset;
+               nread = vfs_read(nef_filp,
+                                (char __user *)flash_memory[i],
+                                GLOB_LLD_PAGE_SIZE, &tmp_file_offset);
+               if (nread < GLOB_LLD_PAGE_SIZE) {
+                       printk(KERN_ERR "%s, Line %d - "
+                              "nand emu file partial read: "
+                              "%d bytes\n", __FILE__, __LINE__, (int)nread);
+                       goto out;
+               }
+               file_offset += GLOB_LLD_PAGE_SIZE;
+       }
+       rc = 0;
+
+out:
+       filp_close(nef_filp, current->files);
+       set_fs(fs);
+       return rc;
+}
+
+/* Write contents of flash_memory to nand emu file */
+int emu_write_mem_to_file(void)
+{
+       mm_segment_t fs;
+       struct file *nef_filp = NULL;
+       struct inode *inode = NULL;
+       loff_t nef_size = 0;
+       loff_t tmp_file_offset, file_offset;
+       ssize_t nwritten;
+       int i, rc = -EINVAL;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       fs = get_fs();
+       set_fs(get_ds());
+
+       nef_filp = filp_open("/root/nand_emu_file", O_RDWR | O_LARGEFILE, 0);
+       if (IS_ERR(nef_filp)) {
+               printk(KERN_ERR "filp_open error: "
+                      "Unable to open nand emu file!\n");
+               return PTR_ERR(nef_filp);
+       }
+
+       if (nef_filp->f_path.dentry) {
+               inode = nef_filp->f_path.dentry->d_inode;
+       } else {
+               printk(KERN_ERR "Invalid " "nef_filp->f_path.dentry value!\n");
+               goto out;
+       }
+
+       nef_size = i_size_read(inode->i_mapping->host);
+       if (nef_size <= 0) {
+               printk(KERN_ERR "Invalid "
+                      "nand emu file size: 0x%llx\n", nef_size);
+               goto out;
+       } else {
+               nand_dbg_print(NAND_DBG_DEBUG, "nand emu file size: "
+                              "%lld\n", nef_size);
+       }
+
+       file_offset = 0;
+       for (i = 0; i < GLOB_LLD_BLOCKS * GLOB_LLD_PAGES; i++) {
+               tmp_file_offset = file_offset;
+               nwritten = vfs_write(nef_filp,
+                                    (char __user *)flash_memory[i],
+                                    GLOB_LLD_PAGE_SIZE, &tmp_file_offset);
+               if (nwritten < GLOB_LLD_PAGE_SIZE) {
+                       printk(KERN_ERR "%s, Line %d - "
+                              "nand emu file partial write: "
+                              "%d bytes\n", __FILE__, __LINE__, (int)nwritten);
+                       goto out;
+               }
+               file_offset += GLOB_LLD_PAGE_SIZE;
+       }
+       rc = 0;
+
+out:
+       filp_close(nef_filp, current->files);
+       set_fs(fs);
+       return rc;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Flash_Init
+* Inputs:       none
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:  Creates & initializes the flash RAM array.
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Flash_Init(void)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       flash_memory[0] = (u8 *)vmalloc(GLOB_LLD_PAGE_SIZE *
+                                                  GLOB_LLD_BLOCKS *
+                                                  GLOB_LLD_PAGES *
+                                                  sizeof(u8));
+       if (!flash_memory[0]) {
+               printk(KERN_ERR "Fail to allocate memory "
+                      "for nand emulator!\n");
+               return ERR;
+       }
+
+       memset((char *)(flash_memory[0]), 0xFF,
+              GLOB_LLD_PAGE_SIZE * GLOB_LLD_BLOCKS * GLOB_LLD_PAGES *
+              sizeof(u8));
+
+       for (i = 1; i < GLOB_LLD_BLOCKS * GLOB_LLD_PAGES; i++)
+               flash_memory[i] = flash_memory[i - 1] + GLOB_LLD_PAGE_SIZE;
+
+       emu_load_file_to_mem(); /* Load nand emu file to mem */
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Flash_Release
+* Inputs:       none
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Releases the flash.
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+int emu_Flash_Release(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       emu_write_mem_to_file();  /* Write back mem to nand emu file */
+
+       vfree(flash_memory[0]);
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Read_Device_ID
+* Inputs:       none
+* Outputs:      PASS=1 FAIL=0
+* Description:  Reads the info from the controller registers.
+*               Sets up DeviceInfo structure with device parameters
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+
+u16 emu_Read_Device_ID(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       DeviceInfo.wDeviceMaker = 0;
+       DeviceInfo.wDeviceType = 8;
+       DeviceInfo.wSpectraStartBlock = 36;
+       DeviceInfo.wSpectraEndBlock = GLOB_LLD_BLOCKS - 1;
+       DeviceInfo.wTotalBlocks = GLOB_LLD_BLOCKS;
+       DeviceInfo.wPagesPerBlock = GLOB_LLD_PAGES;
+       DeviceInfo.wPageSize = GLOB_LLD_PAGE_SIZE;
+       DeviceInfo.wPageDataSize = GLOB_LLD_PAGE_DATA_SIZE;
+       DeviceInfo.wPageSpareSize = GLOB_LLD_PAGE_SIZE -
+           GLOB_LLD_PAGE_DATA_SIZE;
+       DeviceInfo.wBlockSize = DeviceInfo.wPageSize * GLOB_LLD_PAGES;
+       DeviceInfo.wBlockDataSize = DeviceInfo.wPageDataSize * GLOB_LLD_PAGES;
+       DeviceInfo.wDataBlockNum = (u32) (DeviceInfo.wSpectraEndBlock -
+                                               DeviceInfo.wSpectraStartBlock
+                                               + 1);
+       DeviceInfo.MLCDevice = 1; /* Emulate MLC device */
+       DeviceInfo.nBitsInPageNumber =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wPagesPerBlock);
+       DeviceInfo.nBitsInPageDataSize =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wPageDataSize);
+       DeviceInfo.nBitsInBlockDataSize =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wBlockDataSize);
+
+#if CMD_DMA
+       totalUsedBanks = 4;
+       valid_banks[0] = 1;
+       valid_banks[1] = 1;
+       valid_banks[2] = 1;
+       valid_banks[3] = 1;
+#endif
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Flash_Reset
+* Inputs:       none
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Reset the flash
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Flash_Reset(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Erase_Block
+* Inputs:       Address
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Erase a block
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Erase_Block(u32 block_add)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (block_add >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "emu_Erase_Block error! "
+                      "Too big block address: %d\n", block_add);
+               return FAIL;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Erasing block %d\n",
+               (int)block_add);
+
+       for (i = block_add * GLOB_LLD_PAGES;
+            i < ((block_add + 1) * GLOB_LLD_PAGES); i++) {
+               if (flash_memory[i]) {
+                       memset((u8 *)(flash_memory[i]), 0xFF,
+                              DeviceInfo.wPageSize * sizeof(u8));
+               }
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Write_Page_Main
+* Inputs:       Write buffer address pointer
+*               Block number
+*               Page  number
+*               Number of pages to process
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:  Write the data in the buffer to main area of flash
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Write_Page_Main(u8 *write_data, u32 Block,
+                          u16 Page, u16 PageCount)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks)
+               return FAIL;
+
+       if (Page + PageCount > DeviceInfo.wPagesPerBlock)
+               return FAIL;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "emu_Write_Page_Main: "
+                      "lba %u Page %u PageCount %u\n",
+                      (unsigned int)Block,
+                      (unsigned int)Page, (unsigned int)PageCount);
+
+       for (i = 0; i < PageCount; i++) {
+               if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+                       printk(KERN_ERR "Run out of memory\n");
+                       return FAIL;
+               }
+               memcpy((u8 *) (flash_memory[Block * GLOB_LLD_PAGES + Page]),
+                      write_data, DeviceInfo.wPageDataSize);
+               write_data += DeviceInfo.wPageDataSize;
+               Page++;
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Read_Page_Main
+* Inputs:       Read buffer address pointer
+*               Block number
+*               Page  number
+*               Number of pages to process
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:  Read the data from the flash main area to the buffer
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Read_Page_Main(u8 *read_data, u32 Block,
+                         u16 Page, u16 PageCount)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks)
+               return FAIL;
+
+       if (Page + PageCount > DeviceInfo.wPagesPerBlock)
+               return FAIL;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "emu_Read_Page_Main: "
+                      "lba %u Page %u PageCount %u\n",
+                      (unsigned int)Block,
+                      (unsigned int)Page, (unsigned int)PageCount);
+
+       for (i = 0; i < PageCount; i++) {
+               if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+                       memset(read_data, 0xFF, DeviceInfo.wPageDataSize);
+               } else {
+                       memcpy(read_data,
+                              (u8 *) (flash_memory[Block * GLOB_LLD_PAGES
+                                                     + Page]),
+                              DeviceInfo.wPageDataSize);
+               }
+               read_data += DeviceInfo.wPageDataSize;
+               Page++;
+       }
+
+       return PASS;
+}
+
+#ifndef ELDORA
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Read_Page_Main_Spare
+* Inputs:       Write Buffer
+*                       Address
+*                       Buffer size
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Read from flash main+spare area
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Read_Page_Main_Spare(u8 *read_data, u32 Block,
+                               u16 Page, u16 PageCount)
+{
+       int i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "Read Page Main+Spare "
+                      "Error: Block Address too big\n");
+               return FAIL;
+       }
+
+       if (Page + PageCount > DeviceInfo.wPagesPerBlock) {
+               printk(KERN_ERR "Read Page Main+Spare "
+                      "Error: Page number too big\n");
+               return FAIL;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Read Page Main + Spare - "
+                      "No. of pages %u block %u start page %u\n",
+                      (unsigned int)PageCount,
+                      (unsigned int)Block, (unsigned int)Page);
+
+       for (i = 0; i < PageCount; i++) {
+               if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+                       memset(read_data, 0xFF, DeviceInfo.wPageSize);
+               } else {
+                       memcpy(read_data, (u8 *) (flash_memory[Block *
+                                                                GLOB_LLD_PAGES
+                                                                + Page]),
+                              DeviceInfo.wPageSize);
+               }
+
+               read_data += DeviceInfo.wPageSize;
+               Page++;
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Write_Page_Main_Spare
+* Inputs:       Write buffer
+*                       address
+*                       buffer length
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Write the buffer to main+spare area of flash
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Write_Page_Main_Spare(u8 *write_data, u32 Block,
+                                u16 Page, u16 page_count)
+{
+       u16 i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "Write Page Main + Spare "
+                      "Error: Block Address too big\n");
+               return FAIL;
+       }
+
+       if (Page + page_count > DeviceInfo.wPagesPerBlock) {
+               printk(KERN_ERR "Write Page Main + Spare "
+                      "Error: Page number too big\n");
+               return FAIL;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Write Page Main+Spare - "
+                      "No. of pages %u block %u start page %u\n",
+                      (unsigned int)page_count,
+                      (unsigned int)Block, (unsigned int)Page);
+
+       for (i = 0; i < page_count; i++) {
+               if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+                       printk(KERN_ERR "Run out of memory!\n");
+                       return FAIL;
+               }
+               memcpy((u8 *) (flash_memory[Block * GLOB_LLD_PAGES + Page]),
+                      write_data, DeviceInfo.wPageSize);
+               write_data += DeviceInfo.wPageSize;
+               Page++;
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Write_Page_Spare
+* Inputs:       Write buffer
+*                       Address
+*                       buffer size
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Write the buffer in the spare area
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Write_Page_Spare(u8 *write_data, u32 Block,
+                           u16 Page, u16 PageCount)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "Read Page Spare Error: "
+                      "Block Address too big\n");
+               return FAIL;
+       }
+
+       if (Page + PageCount > DeviceInfo.wPagesPerBlock) {
+               printk(KERN_ERR "Read Page Spare Error: "
+                      "Page number too big\n");
+               return FAIL;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Write Page Spare- "
+                      "block %u page %u\n",
+                      (unsigned int)Block, (unsigned int)Page);
+
+       if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+               printk(KERN_ERR "Run out of memory!\n");
+               return FAIL;
+       }
+
+       memcpy((u8 *) (flash_memory[Block * GLOB_LLD_PAGES + Page] +
+                        DeviceInfo.wPageDataSize), write_data,
+              (DeviceInfo.wPageSize - DeviceInfo.wPageDataSize));
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Read_Page_Spare
+* Inputs:       Write Buffer
+*                       Address
+*                       Buffer size
+* Outputs:      PASS=0 (notice 0=ok here)
+* Description:          Read data from the spare area
+*
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_Read_Page_Spare(u8 *write_data, u32 Block,
+                          u16 Page, u16 PageCount)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (Block >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "Read Page Spare "
+                      "Error: Block Address too big\n");
+               return FAIL;
+       }
+
+       if (Page + PageCount > DeviceInfo.wPagesPerBlock) {
+               printk(KERN_ERR "Read Page Spare "
+                      "Error: Page number too big\n");
+               return FAIL;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Read Page Spare- "
+                      "block %u page %u\n",
+                      (unsigned int)Block, (unsigned int)Page);
+
+       if (NULL == flash_memory[Block * GLOB_LLD_PAGES + Page]) {
+               memset(write_data, 0xFF,
+                      (DeviceInfo.wPageSize - DeviceInfo.wPageDataSize));
+       } else {
+               memcpy(write_data,
+                      (u8 *) (flash_memory[Block * GLOB_LLD_PAGES + Page]
+                                + DeviceInfo.wPageDataSize),
+                      (DeviceInfo.wPageSize - DeviceInfo.wPageDataSize));
+       }
+
+       return PASS;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Enable_Disable_Interrupts
+* Inputs:       enable or disable
+* Outputs:      none
+* Description:  NOP
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+void emu_Enable_Disable_Interrupts(u16 INT_ENABLE)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+}
+
+u16 emu_Get_Bad_Block(u32 block)
+{
+       return 0;
+}
+
+#if CMD_DMA
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Support for CDMA functions
+************************************
+*       emu_CDMA_Flash_Init
+*           CDMA_process_data command   (use LLD_CDMA)
+*           CDMA_MemCopy_CMD            (use LLD_CDMA)
+*       emu_CDMA_execute all commands
+*       emu_CDMA_Event_Status
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_CDMA_Flash_Init(void)
+{
+       u16 i;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0; i < MAX_DESCS + MAX_CHANS; i++) {
+               PendingCMD[i].CMD = 0;
+               PendingCMD[i].Tag = 0;
+               PendingCMD[i].DataAddr = 0;
+               PendingCMD[i].Block = 0;
+               PendingCMD[i].Page = 0;
+               PendingCMD[i].PageCount = 0;
+               PendingCMD[i].DataDestAddr = 0;
+               PendingCMD[i].DataSrcAddr = 0;
+               PendingCMD[i].MemCopyByteCnt = 0;
+               PendingCMD[i].ChanSync[0] = 0;
+               PendingCMD[i].ChanSync[1] = 0;
+               PendingCMD[i].ChanSync[2] = 0;
+               PendingCMD[i].ChanSync[3] = 0;
+               PendingCMD[i].ChanSync[4] = 0;
+               PendingCMD[i].Status = 3;
+       }
+
+       return PASS;
+}
+
+static void emu_isr(int irq, void *dev_id)
+{
+       /* TODO:  ... */
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     CDMA_Execute_CMDs
+* Inputs:       tag_count:  the number of pending cmds to do
+* Outputs:      PASS/FAIL
+* Description:  execute each command in the pending CMD array
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_CDMA_Execute_CMDs(u16 tag_count)
+{
+       u16 i, j;
+       u8 CMD;         /* cmd parameter */
+       u8 *data;
+       u32 block;
+       u16 page;
+       u16 count;
+       u16 status = PASS;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       nand_dbg_print(NAND_DBG_TRACE, "At start of Execute CMDs: "
+                      "Tag Count %u\n", tag_count);
+
+       for (i = 0; i < totalUsedBanks; i++) {
+               PendingCMD[i].CMD = DUMMY_CMD;
+               PendingCMD[i].Tag = 0xFF;
+               PendingCMD[i].Block =
+                   (DeviceInfo.wTotalBlocks / totalUsedBanks) * i;
+
+               for (j = 0; j <= MAX_CHANS; j++)
+                       PendingCMD[i].ChanSync[j] = 0;
+       }
+
+       CDMA_Execute_CMDs(tag_count);
+
+       print_pending_cmds(tag_count);
+
+#if DEBUG_SYNC
+       }
+       debug_sync_cnt++;
+#endif
+
+       for (i = MAX_CHANS;
+            i < tag_count + MAX_CHANS; i++) {
+               CMD = PendingCMD[i].CMD;
+               data = PendingCMD[i].DataAddr;
+               block = PendingCMD[i].Block;
+               page = PendingCMD[i].Page;
+               count = PendingCMD[i].PageCount;
+
+               switch (CMD) {
+               case ERASE_CMD:
+                       emu_Erase_Block(block);
+                       PendingCMD[i].Status = PASS;
+                       break;
+               case WRITE_MAIN_CMD:
+                       emu_Write_Page_Main(data, block, page, count);
+                       PendingCMD[i].Status = PASS;
+                       break;
+               case WRITE_MAIN_SPARE_CMD:
+                       emu_Write_Page_Main_Spare(data, block, page, count);
+                       PendingCMD[i].Status = PASS;
+                       break;
+               case READ_MAIN_CMD:
+                       emu_Read_Page_Main(data, block, page, count);
+                       PendingCMD[i].Status = PASS;
+                       break;
+               case MEMCOPY_CMD:
+                       memcpy(PendingCMD[i].DataDestAddr,
+                              PendingCMD[i].DataSrcAddr,
+                              PendingCMD[i].MemCopyByteCnt);
+               case DUMMY_CMD:
+                       PendingCMD[i].Status = PASS;
+                       break;
+               default:
+                       PendingCMD[i].Status = FAIL;
+                       break;
+               }
+       }
+
+       /*
+        * Temperory adding code to reset PendingCMD array for basic testing.
+        * It should be done at the end of  event status function.
+        */
+       for (i = tag_count + MAX_CHANS; i < MAX_DESCS; i++) {
+               PendingCMD[i].CMD = 0;
+               PendingCMD[i].Tag = 0;
+               PendingCMD[i].DataAddr = 0;
+               PendingCMD[i].Block = 0;
+               PendingCMD[i].Page = 0;
+               PendingCMD[i].PageCount = 0;
+               PendingCMD[i].DataDestAddr = 0;
+               PendingCMD[i].DataSrcAddr = 0;
+               PendingCMD[i].MemCopyByteCnt = 0;
+               PendingCMD[i].ChanSync[0] = 0;
+               PendingCMD[i].ChanSync[1] = 0;
+               PendingCMD[i].ChanSync[2] = 0;
+               PendingCMD[i].ChanSync[3] = 0;
+               PendingCMD[i].ChanSync[4] = 0;
+               PendingCMD[i].Status = CMD_NOT_DONE;
+       }
+
+       nand_dbg_print(NAND_DBG_TRACE, "At end of Execute CMDs.\n");
+
+       emu_isr(0, 0); /* This is a null isr now. Need fill it in future */
+
+       return status;
+}
+
+/*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
+* Function:     emu_Event_Status
+* Inputs:       none
+* Outputs:      Event_Status code
+* Description:  This function can also be used to force errors
+*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
+u16 emu_CDMA_Event_Status(void)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       return EVENT_PASS;
+}
+
+#endif /* CMD_DMA */
+#endif /* !ELDORA */
+#endif /* FLASH_EMU */
diff --git a/drivers/block/spectra/lld_emu.h b/drivers/block/spectra/lld_emu.h
new file mode 100644 (file)
index 0000000..63f84c3
--- /dev/null
@@ -0,0 +1,51 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _LLD_EMU_
+#define _LLD_EMU_
+
+#include "ffsport.h"
+#include "ffsdefs.h"
+
+/* prototypes: emulator API functions */
+extern u16 emu_Flash_Reset(void);
+extern u16 emu_Flash_Init(void);
+extern int emu_Flash_Release(void);
+extern u16 emu_Read_Device_ID(void);
+extern u16 emu_Erase_Block(u32 block_addr);
+extern u16 emu_Write_Page_Main(u8 *write_data, u32 Block,
+                               u16 Page, u16 PageCount);
+extern u16 emu_Read_Page_Main(u8 *read_data, u32 Block, u16 Page,
+                                u16 PageCount);
+extern u16 emu_Event_Status(void);
+extern void emu_Enable_Disable_Interrupts(u16 INT_ENABLE);
+extern u16 emu_Write_Page_Main_Spare(u8 *write_data, u32 Block,
+                                       u16 Page, u16 PageCount);
+extern u16 emu_Write_Page_Spare(u8 *write_data, u32 Block,
+                                       u16 Page, u16 PageCount);
+extern u16 emu_Read_Page_Main_Spare(u8 *read_data, u32 Block,
+                                      u16 Page, u16 PageCount);
+extern u16 emu_Read_Page_Spare(u8 *read_data, u32 Block, u16 Page,
+                                 u16 PageCount);
+extern u16 emu_Get_Bad_Block(u32 block);
+
+u16 emu_CDMA_Flash_Init(void);
+u16 emu_CDMA_Execute_CMDs(u16 tag_count);
+u16 emu_CDMA_Event_Status(void);
+#endif /*_LLD_EMU_*/
diff --git a/drivers/block/spectra/lld_nand.c b/drivers/block/spectra/lld_nand.c
new file mode 100644 (file)
index 0000000..8c279b8
--- /dev/null
@@ -0,0 +1,2601 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#include "lld.h"
+#include "lld_nand.h"
+#include "lld_cdma.h"
+
+#include "spectraswconfig.h"
+#include "flash.h"
+#include "ffsdefs.h"
+
+#include <linux/interrupt.h>
+#include <linux/delay.h>
+#include <linux/wait.h>
+#include <linux/mutex.h>
+
+#include "nand_regs.h"
+
+#define SPECTRA_NAND_NAME    "nd"
+
+#define CEIL_DIV(X, Y) (((X)%(Y)) ? ((X)/(Y)+1) : ((X)/(Y)))
+#define MAX_PAGES_PER_RW        128
+
+#define INT_IDLE_STATE                 0
+#define INT_READ_PAGE_MAIN    0x01
+#define INT_WRITE_PAGE_MAIN    0x02
+#define INT_PIPELINE_READ_AHEAD    0x04
+#define INT_PIPELINE_WRITE_AHEAD    0x08
+#define INT_MULTI_PLANE_READ    0x10
+#define INT_MULTI_PLANE_WRITE    0x11
+
+static u32 enable_ecc;
+
+struct mrst_nand_info info;
+
+int totalUsedBanks;
+u32 GLOB_valid_banks[LLD_MAX_FLASH_BANKS];
+
+void __iomem *FlashReg;
+void __iomem *FlashMem;
+
+u16 conf_parameters[] = {
+       0x0000,
+       0x0000,
+       0x01F4,
+       0x01F4,
+       0x01F4,
+       0x01F4,
+       0x0000,
+       0x0000,
+       0x0001,
+       0x0000,
+       0x0000,
+       0x0000,
+       0x0000,
+       0x0040,
+       0x0001,
+       0x000A,
+       0x000A,
+       0x000A,
+       0x0000,
+       0x0000,
+       0x0005,
+       0x0012,
+       0x000C
+};
+
+u16   NAND_Get_Bad_Block(u32 block)
+{
+       u32 status = PASS;
+       u32 flag_bytes  = 0;
+       u32 skip_bytes  = DeviceInfo.wSpareSkipBytes;
+       u32 page, i;
+       u8 *pReadSpareBuf = buf_get_bad_block;
+
+       if (enable_ecc)
+               flag_bytes = DeviceInfo.wNumPageSpareFlag;
+
+       for (page = 0; page < 2; page++) {
+               status = NAND_Read_Page_Spare(pReadSpareBuf, block, page, 1);
+               if (status != PASS)
+                       return READ_ERROR;
+               for (i = flag_bytes; i < (flag_bytes + skip_bytes); i++)
+                       if (pReadSpareBuf[i] != 0xff)
+                               return DEFECTIVE_BLOCK;
+       }
+
+       for (page = 1; page < 3; page++) {
+               status = NAND_Read_Page_Spare(pReadSpareBuf, block,
+                       DeviceInfo.wPagesPerBlock - page , 1);
+               if (status != PASS)
+                       return READ_ERROR;
+               for (i = flag_bytes; i < (flag_bytes + skip_bytes); i++)
+                       if (pReadSpareBuf[i] != 0xff)
+                               return DEFECTIVE_BLOCK;
+       }
+
+       return GOOD_BLOCK;
+}
+
+
+u16 NAND_Flash_Reset(void)
+{
+       u32 i;
+       u32 intr_status_rst_comp[4] = {INTR_STATUS0__RST_COMP,
+               INTR_STATUS1__RST_COMP,
+               INTR_STATUS2__RST_COMP,
+               INTR_STATUS3__RST_COMP};
+       u32 intr_status_time_out[4] = {INTR_STATUS0__TIME_OUT,
+               INTR_STATUS1__TIME_OUT,
+               INTR_STATUS2__TIME_OUT,
+               INTR_STATUS3__TIME_OUT};
+       u32 intr_status[4] = {INTR_STATUS0, INTR_STATUS1,
+               INTR_STATUS2, INTR_STATUS3};
+       u32 device_reset_banks[4] = {DEVICE_RESET__BANK0,
+               DEVICE_RESET__BANK1,
+               DEVICE_RESET__BANK2,
+               DEVICE_RESET__BANK3};
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       for (i = 0 ; i < LLD_MAX_FLASH_BANKS; i++)
+               iowrite32(intr_status_rst_comp[i] | intr_status_time_out[i],
+               FlashReg + intr_status[i]);
+
+       for (i = 0 ; i < LLD_MAX_FLASH_BANKS; i++) {
+               iowrite32(device_reset_banks[i], FlashReg + DEVICE_RESET);
+               while (!(ioread32(FlashReg + intr_status[i]) &
+                       (intr_status_rst_comp[i] | intr_status_time_out[i])))
+                       ;
+               if (ioread32(FlashReg + intr_status[i]) &
+                       intr_status_time_out[i])
+                       nand_dbg_print(NAND_DBG_WARN,
+                       "NAND Reset operation timed out on bank %d\n", i);
+       }
+
+       for (i = 0; i < LLD_MAX_FLASH_BANKS; i++)
+               iowrite32(intr_status_rst_comp[i] | intr_status_time_out[i],
+                       FlashReg + intr_status[i]);
+
+       return PASS;
+}
+
+static void NAND_ONFi_Timing_Mode(u16 mode)
+{
+       u16 Trea[6] = {40, 30, 25, 20, 20, 16};
+       u16 Trp[6] = {50, 25, 17, 15, 12, 10};
+       u16 Treh[6] = {30, 15, 15, 10, 10, 7};
+       u16 Trc[6] = {100, 50, 35, 30, 25, 20};
+       u16 Trhoh[6] = {0, 15, 15, 15, 15, 15};
+       u16 Trloh[6] = {0, 0, 0, 0, 5, 5};
+       u16 Tcea[6] = {100, 45, 30, 25, 25, 25};
+       u16 Tadl[6] = {200, 100, 100, 100, 70, 70};
+       u16 Trhw[6] = {200, 100, 100, 100, 100, 100};
+       u16 Trhz[6] = {200, 100, 100, 100, 100, 100};
+       u16 Twhr[6] = {120, 80, 80, 60, 60, 60};
+       u16 Tcs[6] = {70, 35, 25, 25, 20, 15};
+
+       u16 TclsRising = 1;
+       u16 data_invalid_rhoh, data_invalid_rloh, data_invalid;
+       u16 dv_window = 0;
+       u16 en_lo, en_hi;
+       u16 acc_clks;
+       u16 addr_2_data, re_2_we, re_2_re, we_2_re, cs_cnt;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       en_lo = CEIL_DIV(Trp[mode], CLK_X);
+       en_hi = CEIL_DIV(Treh[mode], CLK_X);
+
+#if ONFI_BLOOM_TIME
+       if ((en_hi * CLK_X) < (Treh[mode] + 2))
+               en_hi++;
+#endif
+
+       if ((en_lo + en_hi) * CLK_X < Trc[mode])
+               en_lo += CEIL_DIV((Trc[mode] - (en_lo + en_hi) * CLK_X), CLK_X);
+
+       if ((en_lo + en_hi) < CLK_MULTI)
+               en_lo += CLK_MULTI - en_lo - en_hi;
+
+       while (dv_window < 8) {
+               data_invalid_rhoh = en_lo * CLK_X + Trhoh[mode];
+
+               data_invalid_rloh = (en_lo + en_hi) * CLK_X + Trloh[mode];
+
+               data_invalid =
+                   data_invalid_rhoh <
+                   data_invalid_rloh ? data_invalid_rhoh : data_invalid_rloh;
+
+               dv_window = data_invalid - Trea[mode];
+
+               if (dv_window < 8)
+                       en_lo++;
+       }
+
+       acc_clks = CEIL_DIV(Trea[mode], CLK_X);
+
+       while (((acc_clks * CLK_X) - Trea[mode]) < 3)
+               acc_clks++;
+
+       if ((data_invalid - acc_clks * CLK_X) < 2)
+               nand_dbg_print(NAND_DBG_WARN, "%s, Line %d: Warning!\n",
+                       __FILE__, __LINE__);
+
+       addr_2_data = CEIL_DIV(Tadl[mode], CLK_X);
+       re_2_we = CEIL_DIV(Trhw[mode], CLK_X);
+       re_2_re = CEIL_DIV(Trhz[mode], CLK_X);
+       we_2_re = CEIL_DIV(Twhr[mode], CLK_X);
+       cs_cnt = CEIL_DIV((Tcs[mode] - Trp[mode]), CLK_X);
+       if (!TclsRising)
+               cs_cnt = CEIL_DIV(Tcs[mode], CLK_X);
+       if (cs_cnt == 0)
+               cs_cnt = 1;
+
+       if (Tcea[mode]) {
+               while (((cs_cnt * CLK_X) + Trea[mode]) < Tcea[mode])
+                       cs_cnt++;
+       }
+
+#if MODE5_WORKAROUND
+       if (mode == 5)
+               acc_clks = 5;
+#endif
+
+       /* Sighting 3462430: Temporary hack for MT29F128G08CJABAWP:B */
+       if ((ioread32(FlashReg + MANUFACTURER_ID) == 0) &&
+               (ioread32(FlashReg + DEVICE_ID) == 0x88))
+               acc_clks = 6;
+
+       iowrite32(acc_clks, FlashReg + ACC_CLKS);
+       iowrite32(re_2_we, FlashReg + RE_2_WE);
+       iowrite32(re_2_re, FlashReg + RE_2_RE);
+       iowrite32(we_2_re, FlashReg + WE_2_RE);
+       iowrite32(addr_2_data, FlashReg + ADDR_2_DATA);
+       iowrite32(en_lo, FlashReg + RDWR_EN_LO_CNT);
+       iowrite32(en_hi, FlashReg + RDWR_EN_HI_CNT);
+       iowrite32(cs_cnt, FlashReg + CS_SETUP_CNT);
+}
+
+static void index_addr(u32 address, u32 data)
+{
+       iowrite32(address, FlashMem);
+       iowrite32(data, FlashMem + 0x10);
+}
+
+static void index_addr_read_data(u32 address, u32 *pdata)
+{
+       iowrite32(address, FlashMem);
+       *pdata = ioread32(FlashMem + 0x10);
+}
+
+static void set_ecc_config(void)
+{
+#if SUPPORT_8BITECC
+       if ((ioread32(FlashReg + DEVICE_MAIN_AREA_SIZE) < 4096) ||
+               (ioread32(FlashReg + DEVICE_SPARE_AREA_SIZE) <= 128))
+               iowrite32(8, FlashReg + ECC_CORRECTION);
+#endif
+
+       if ((ioread32(FlashReg + ECC_CORRECTION) & ECC_CORRECTION__VALUE)
+               == 1) {
+               DeviceInfo.wECCBytesPerSector = 4;
+               DeviceInfo.wECCBytesPerSector *= DeviceInfo.wDevicesConnected;
+               DeviceInfo.wNumPageSpareFlag =
+                       DeviceInfo.wPageSpareSize -
+                       DeviceInfo.wPageDataSize /
+                       (ECC_SECTOR_SIZE * DeviceInfo.wDevicesConnected) *
+                       DeviceInfo.wECCBytesPerSector
+                       - DeviceInfo.wSpareSkipBytes;
+       } else {
+               DeviceInfo.wECCBytesPerSector =
+                       (ioread32(FlashReg + ECC_CORRECTION) &
+                       ECC_CORRECTION__VALUE) * 13 / 8;
+               if ((DeviceInfo.wECCBytesPerSector) % 2 == 0)
+                       DeviceInfo.wECCBytesPerSector += 2;
+               else
+                       DeviceInfo.wECCBytesPerSector += 1;
+
+               DeviceInfo.wECCBytesPerSector *= DeviceInfo.wDevicesConnected;
+               DeviceInfo.wNumPageSpareFlag = DeviceInfo.wPageSpareSize -
+                       DeviceInfo.wPageDataSize /
+                       (ECC_SECTOR_SIZE * DeviceInfo.wDevicesConnected) *
+                       DeviceInfo.wECCBytesPerSector
+                       - DeviceInfo.wSpareSkipBytes;
+       }
+}
+
+static u16 get_onfi_nand_para(void)
+{
+       int i;
+       u16 blks_lun_l, blks_lun_h, n_of_luns;
+       u32 blockperlun, id;
+
+       iowrite32(DEVICE_RESET__BANK0, FlashReg + DEVICE_RESET);
+
+       while (!((ioread32(FlashReg + INTR_STATUS0) &
+               INTR_STATUS0__RST_COMP) |
+               (ioread32(FlashReg + INTR_STATUS0) &
+               INTR_STATUS0__TIME_OUT)))
+               ;
+
+       if (ioread32(FlashReg + INTR_STATUS0) & INTR_STATUS0__RST_COMP) {
+               iowrite32(DEVICE_RESET__BANK1, FlashReg + DEVICE_RESET);
+               while (!((ioread32(FlashReg + INTR_STATUS1) &
+                       INTR_STATUS1__RST_COMP) |
+                       (ioread32(FlashReg + INTR_STATUS1) &
+                       INTR_STATUS1__TIME_OUT)))
+                       ;
+
+               if (ioread32(FlashReg + INTR_STATUS1) &
+                       INTR_STATUS1__RST_COMP) {
+                       iowrite32(DEVICE_RESET__BANK2,
+                               FlashReg + DEVICE_RESET);
+                       while (!((ioread32(FlashReg + INTR_STATUS2) &
+                               INTR_STATUS2__RST_COMP) |
+                               (ioread32(FlashReg + INTR_STATUS2) &
+                               INTR_STATUS2__TIME_OUT)))
+                               ;
+
+                       if (ioread32(FlashReg + INTR_STATUS2) &
+                               INTR_STATUS2__RST_COMP) {
+                               iowrite32(DEVICE_RESET__BANK3,
+                                       FlashReg + DEVICE_RESET);
+                               while (!((ioread32(FlashReg + INTR_STATUS3) &
+                                       INTR_STATUS3__RST_COMP) |
+                                       (ioread32(FlashReg + INTR_STATUS3) &
+                                       INTR_STATUS3__TIME_OUT)))
+                                       ;
+                       } else {
+                               printk(KERN_ERR "Getting a time out for bank 2!\n");
+                       }
+               } else {
+                       printk(KERN_ERR "Getting a time out for bank 1!\n");
+               }
+       }
+
+       iowrite32(INTR_STATUS0__TIME_OUT, FlashReg + INTR_STATUS0);
+       iowrite32(INTR_STATUS1__TIME_OUT, FlashReg + INTR_STATUS1);
+       iowrite32(INTR_STATUS2__TIME_OUT, FlashReg + INTR_STATUS2);
+       iowrite32(INTR_STATUS3__TIME_OUT, FlashReg + INTR_STATUS3);
+
+       DeviceInfo.wONFIDevFeatures =
+               ioread32(FlashReg + ONFI_DEVICE_FEATURES);
+       DeviceInfo.wONFIOptCommands =
+               ioread32(FlashReg + ONFI_OPTIONAL_COMMANDS);
+       DeviceInfo.wONFITimingMode =
+               ioread32(FlashReg + ONFI_TIMING_MODE);
+       DeviceInfo.wONFIPgmCacheTimingMode =
+               ioread32(FlashReg + ONFI_PGM_CACHE_TIMING_MODE);
+
+       n_of_luns = ioread32(FlashReg + ONFI_DEVICE_NO_OF_LUNS) &
+               ONFI_DEVICE_NO_OF_LUNS__NO_OF_LUNS;
+       blks_lun_l = ioread32(FlashReg + ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_L);
+       blks_lun_h = ioread32(FlashReg + ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_U);
+
+       blockperlun = (blks_lun_h << 16) | blks_lun_l;
+
+       DeviceInfo.wTotalBlocks = n_of_luns * blockperlun;
+
+       if (!(ioread32(FlashReg + ONFI_TIMING_MODE) &
+               ONFI_TIMING_MODE__VALUE))
+               return FAIL;
+
+       for (i = 5; i > 0; i--) {
+               if (ioread32(FlashReg + ONFI_TIMING_MODE) & (0x01 << i))
+                       break;
+       }
+
+       NAND_ONFi_Timing_Mode(i);
+
+       index_addr(MODE_11 | 0, 0x90);
+       index_addr(MODE_11 | 1, 0);
+
+       for (i = 0; i < 3; i++)
+               index_addr_read_data(MODE_11 | 2, &id);
+
+       nand_dbg_print(NAND_DBG_DEBUG, "3rd ID: 0x%x\n", id);
+
+       DeviceInfo.MLCDevice = id & 0x0C;
+
+       /* By now, all the ONFI devices we know support the page cache */
+       /* rw feature. So here we enable the pipeline_rw_ahead feature */
+       /* iowrite32(1, FlashReg + CACHE_WRITE_ENABLE); */
+       /* iowrite32(1, FlashReg + CACHE_READ_ENABLE);  */
+
+       return PASS;
+}
+
+static void get_samsung_nand_para(void)
+{
+       u8 no_of_planes;
+       u32 blk_size;
+       u64 plane_size, capacity;
+       u32 id_bytes[5];
+       int i;
+
+       index_addr((u32)(MODE_11 | 0), 0x90);
+       index_addr((u32)(MODE_11 | 1), 0);
+       for (i = 0; i < 5; i++)
+               index_addr_read_data((u32)(MODE_11 | 2), &id_bytes[i]);
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "ID bytes: 0x%x, 0x%x, 0x%x, 0x%x, 0x%x\n",
+               id_bytes[0], id_bytes[1], id_bytes[2],
+               id_bytes[3], id_bytes[4]);
+
+       if ((id_bytes[1] & 0xff) == 0xd3) { /* Samsung K9WAG08U1A */
+               /* Set timing register values according to datasheet */
+               iowrite32(5, FlashReg + ACC_CLKS);
+               iowrite32(20, FlashReg + RE_2_WE);
+               iowrite32(12, FlashReg + WE_2_RE);
+               iowrite32(14, FlashReg + ADDR_2_DATA);
+               iowrite32(3, FlashReg + RDWR_EN_LO_CNT);
+               iowrite32(2, FlashReg + RDWR_EN_HI_CNT);
+               iowrite32(2, FlashReg + CS_SETUP_CNT);
+       }
+
+       no_of_planes = 1 << ((id_bytes[4] & 0x0c) >> 2);
+       plane_size  = (u64)64 << ((id_bytes[4] & 0x70) >> 4);
+       blk_size = 64 << ((ioread32(FlashReg + DEVICE_PARAM_1) & 0x30) >> 4);
+       capacity = (u64)128 * plane_size * no_of_planes;
+
+       DeviceInfo.wTotalBlocks = (u32)GLOB_u64_Div(capacity, blk_size);
+}
+
+static void get_toshiba_nand_para(void)
+{
+       void __iomem *scratch_reg;
+       u32 tmp;
+
+       /* Workaround to fix a controller bug which reports a wrong */
+       /* spare area size for some kind of Toshiba NAND device */
+       if ((ioread32(FlashReg + DEVICE_MAIN_AREA_SIZE) == 4096) &&
+               (ioread32(FlashReg + DEVICE_SPARE_AREA_SIZE) == 64)) {
+               iowrite32(216, FlashReg + DEVICE_SPARE_AREA_SIZE);
+               tmp = ioread32(FlashReg + DEVICES_CONNECTED) *
+                       ioread32(FlashReg + DEVICE_SPARE_AREA_SIZE);
+               iowrite32(tmp, FlashReg + LOGICAL_PAGE_SPARE_SIZE);
+#if SUPPORT_15BITECC
+               iowrite32(15, FlashReg + ECC_CORRECTION);
+#elif SUPPORT_8BITECC
+               iowrite32(8, FlashReg + ECC_CORRECTION);
+#endif
+       }
+
+       /* As Toshiba NAND can not provide it's block number, */
+       /* so here we need user to provide the correct block */
+       /* number in a scratch register before the Linux NAND */
+       /* driver is loaded. If no valid value found in the scratch */
+       /* register, then we use default block number value */
+       scratch_reg = ioremap_nocache(SCRATCH_REG_ADDR, SCRATCH_REG_SIZE);
+       if (!scratch_reg) {
+               printk(KERN_ERR "Spectra: ioremap failed in %s, Line %d",
+                       __FILE__, __LINE__);
+               DeviceInfo.wTotalBlocks = GLOB_HWCTL_DEFAULT_BLKS;
+       } else {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Spectra: ioremap reg address: 0x%p\n", scratch_reg);
+               DeviceInfo.wTotalBlocks = 1 << ioread8(scratch_reg);
+               if (DeviceInfo.wTotalBlocks < 512)
+                       DeviceInfo.wTotalBlocks = GLOB_HWCTL_DEFAULT_BLKS;
+               iounmap(scratch_reg);
+       }
+}
+
+static void get_hynix_nand_para(void)
+{
+       void __iomem *scratch_reg;
+       u32 main_size, spare_size;
+
+       switch (DeviceInfo.wDeviceID) {
+       case 0xD5: /* Hynix H27UAG8T2A, H27UBG8U5A or H27UCG8VFA */
+       case 0xD7: /* Hynix H27UDG8VEM, H27UCG8UDM or H27UCG8V5A */
+               iowrite32(128, FlashReg + PAGES_PER_BLOCK);
+               iowrite32(4096, FlashReg + DEVICE_MAIN_AREA_SIZE);
+               iowrite32(224, FlashReg + DEVICE_SPARE_AREA_SIZE);
+               main_size = 4096 * ioread32(FlashReg + DEVICES_CONNECTED);
+               spare_size = 224 * ioread32(FlashReg + DEVICES_CONNECTED);
+               iowrite32(main_size, FlashReg + LOGICAL_PAGE_DATA_SIZE);
+               iowrite32(spare_size, FlashReg + LOGICAL_PAGE_SPARE_SIZE);
+               iowrite32(0, FlashReg + DEVICE_WIDTH);
+#if SUPPORT_15BITECC
+               iowrite32(15, FlashReg + ECC_CORRECTION);
+#elif SUPPORT_8BITECC
+               iowrite32(8, FlashReg + ECC_CORRECTION);
+#endif
+               DeviceInfo.MLCDevice  = 1;
+               break;
+       default:
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Spectra: Unknown Hynix NAND (Device ID: 0x%x)."
+                       "Will use default parameter values instead.\n",
+                       DeviceInfo.wDeviceID);
+       }
+
+       scratch_reg = ioremap_nocache(SCRATCH_REG_ADDR, SCRATCH_REG_SIZE);
+       if (!scratch_reg) {
+               printk(KERN_ERR "Spectra: ioremap failed in %s, Line %d",
+                       __FILE__, __LINE__);
+               DeviceInfo.wTotalBlocks = GLOB_HWCTL_DEFAULT_BLKS;
+       } else {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Spectra: ioremap reg address: 0x%p\n", scratch_reg);
+               DeviceInfo.wTotalBlocks = 1 << ioread8(scratch_reg);
+               if (DeviceInfo.wTotalBlocks < 512)
+                       DeviceInfo.wTotalBlocks = GLOB_HWCTL_DEFAULT_BLKS;
+               iounmap(scratch_reg);
+       }
+}
+
+static void find_valid_banks(void)
+{
+       u32 id[LLD_MAX_FLASH_BANKS];
+       int i;
+
+       totalUsedBanks = 0;
+       for (i = 0; i < LLD_MAX_FLASH_BANKS; i++) {
+               index_addr((u32)(MODE_11 | (i << 24) | 0), 0x90);
+               index_addr((u32)(MODE_11 | (i << 24) | 1), 0);
+               index_addr_read_data((u32)(MODE_11 | (i << 24) | 2), &id[i]);
+
+               nand_dbg_print(NAND_DBG_DEBUG,
+                       "Return 1st ID for bank[%d]: %x\n", i, id[i]);
+
+               if (i == 0) {
+                       if (id[i] & 0x0ff)
+                               GLOB_valid_banks[i] = 1;
+               } else {
+                       if ((id[i] & 0x0ff) == (id[0] & 0x0ff))
+                               GLOB_valid_banks[i] = 1;
+               }
+
+               totalUsedBanks += GLOB_valid_banks[i];
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "totalUsedBanks: %d\n", totalUsedBanks);
+}
+
+static void detect_partition_feature(void)
+{
+       if (ioread32(FlashReg + FEATURES) & FEATURES__PARTITION) {
+               if ((ioread32(FlashReg + PERM_SRC_ID_1) &
+                       PERM_SRC_ID_1__SRCID) == SPECTRA_PARTITION_ID) {
+                       DeviceInfo.wSpectraStartBlock =
+                           ((ioread32(FlashReg + MIN_MAX_BANK_1) &
+                             MIN_MAX_BANK_1__MIN_VALUE) *
+                            DeviceInfo.wTotalBlocks)
+                           +
+                           (ioread32(FlashReg + MIN_BLK_ADDR_1) &
+                           MIN_BLK_ADDR_1__VALUE);
+
+                       DeviceInfo.wSpectraEndBlock =
+                           (((ioread32(FlashReg + MIN_MAX_BANK_1) &
+                              MIN_MAX_BANK_1__MAX_VALUE) >> 2) *
+                            DeviceInfo.wTotalBlocks)
+                           +
+                           (ioread32(FlashReg + MAX_BLK_ADDR_1) &
+                           MAX_BLK_ADDR_1__VALUE);
+
+                       DeviceInfo.wTotalBlocks *= totalUsedBanks;
+
+                       if (DeviceInfo.wSpectraEndBlock >=
+                           DeviceInfo.wTotalBlocks) {
+                               DeviceInfo.wSpectraEndBlock =
+                                   DeviceInfo.wTotalBlocks - 1;
+                       }
+
+                       DeviceInfo.wDataBlockNum =
+                               DeviceInfo.wSpectraEndBlock -
+                               DeviceInfo.wSpectraStartBlock + 1;
+               } else {
+                       DeviceInfo.wTotalBlocks *= totalUsedBanks;
+                       DeviceInfo.wSpectraStartBlock = SPECTRA_START_BLOCK;
+                       DeviceInfo.wSpectraEndBlock =
+                               DeviceInfo.wTotalBlocks - 1;
+                       DeviceInfo.wDataBlockNum =
+                               DeviceInfo.wSpectraEndBlock -
+                               DeviceInfo.wSpectraStartBlock + 1;
+               }
+       } else {
+               DeviceInfo.wTotalBlocks *= totalUsedBanks;
+               DeviceInfo.wSpectraStartBlock = SPECTRA_START_BLOCK;
+               DeviceInfo.wSpectraEndBlock = DeviceInfo.wTotalBlocks - 1;
+               DeviceInfo.wDataBlockNum =
+                       DeviceInfo.wSpectraEndBlock -
+                       DeviceInfo.wSpectraStartBlock + 1;
+       }
+}
+
+static void dump_device_info(void)
+{
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceInfo:\n");
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceMaker: 0x%x\n",
+               DeviceInfo.wDeviceMaker);
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceID: 0x%x\n",
+               DeviceInfo.wDeviceID);
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceType: 0x%x\n",
+               DeviceInfo.wDeviceType);
+       nand_dbg_print(NAND_DBG_DEBUG, "SpectraStartBlock: %d\n",
+               DeviceInfo.wSpectraStartBlock);
+       nand_dbg_print(NAND_DBG_DEBUG, "SpectraEndBlock: %d\n",
+               DeviceInfo.wSpectraEndBlock);
+       nand_dbg_print(NAND_DBG_DEBUG, "TotalBlocks: %d\n",
+               DeviceInfo.wTotalBlocks);
+       nand_dbg_print(NAND_DBG_DEBUG, "PagesPerBlock: %d\n",
+               DeviceInfo.wPagesPerBlock);
+       nand_dbg_print(NAND_DBG_DEBUG, "PageSize: %d\n",
+               DeviceInfo.wPageSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "PageDataSize: %d\n",
+               DeviceInfo.wPageDataSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "PageSpareSize: %d\n",
+               DeviceInfo.wPageSpareSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "NumPageSpareFlag: %d\n",
+               DeviceInfo.wNumPageSpareFlag);
+       nand_dbg_print(NAND_DBG_DEBUG, "ECCBytesPerSector: %d\n",
+               DeviceInfo.wECCBytesPerSector);
+       nand_dbg_print(NAND_DBG_DEBUG, "BlockSize: %d\n",
+               DeviceInfo.wBlockSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "BlockDataSize: %d\n",
+               DeviceInfo.wBlockDataSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "DataBlockNum: %d\n",
+               DeviceInfo.wDataBlockNum);
+       nand_dbg_print(NAND_DBG_DEBUG, "PlaneNum: %d\n",
+               DeviceInfo.bPlaneNum);
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceMainAreaSize: %d\n",
+               DeviceInfo.wDeviceMainAreaSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceSpareAreaSize: %d\n",
+               DeviceInfo.wDeviceSpareAreaSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "DevicesConnected: %d\n",
+               DeviceInfo.wDevicesConnected);
+       nand_dbg_print(NAND_DBG_DEBUG, "DeviceWidth: %d\n",
+               DeviceInfo.wDeviceWidth);
+       nand_dbg_print(NAND_DBG_DEBUG, "HWRevision: 0x%x\n",
+               DeviceInfo.wHWRevision);
+       nand_dbg_print(NAND_DBG_DEBUG, "HWFeatures: 0x%x\n",
+               DeviceInfo.wHWFeatures);
+       nand_dbg_print(NAND_DBG_DEBUG, "ONFIDevFeatures: 0x%x\n",
+               DeviceInfo.wONFIDevFeatures);
+       nand_dbg_print(NAND_DBG_DEBUG, "ONFIOptCommands: 0x%x\n",
+               DeviceInfo.wONFIOptCommands);
+       nand_dbg_print(NAND_DBG_DEBUG, "ONFITimingMode: 0x%x\n",
+               DeviceInfo.wONFITimingMode);
+       nand_dbg_print(NAND_DBG_DEBUG, "ONFIPgmCacheTimingMode: 0x%x\n",
+               DeviceInfo.wONFIPgmCacheTimingMode);
+       nand_dbg_print(NAND_DBG_DEBUG, "MLCDevice: %s\n",
+               DeviceInfo.MLCDevice ? "Yes" : "No");
+       nand_dbg_print(NAND_DBG_DEBUG, "SpareSkipBytes: %d\n",
+               DeviceInfo.wSpareSkipBytes);
+       nand_dbg_print(NAND_DBG_DEBUG, "BitsInPageNumber: %d\n",
+               DeviceInfo.nBitsInPageNumber);
+       nand_dbg_print(NAND_DBG_DEBUG, "BitsInPageDataSize: %d\n",
+               DeviceInfo.nBitsInPageDataSize);
+       nand_dbg_print(NAND_DBG_DEBUG, "BitsInBlockDataSize: %d\n",
+               DeviceInfo.nBitsInBlockDataSize);
+}
+
+u16 NAND_Read_Device_ID(void)
+{
+       u16 status = PASS;
+       u8 no_of_planes;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       iowrite32(0x02, FlashReg + SPARE_AREA_SKIP_BYTES);
+       iowrite32(0xffff, FlashReg + SPARE_AREA_MARKER);
+       DeviceInfo.wDeviceMaker = ioread32(FlashReg + MANUFACTURER_ID);
+       DeviceInfo.wDeviceID = ioread32(FlashReg + DEVICE_ID);
+       DeviceInfo.MLCDevice = ioread32(FlashReg + DEVICE_PARAM_0) & 0x0c;
+
+       if (ioread32(FlashReg + ONFI_DEVICE_NO_OF_LUNS) &
+               ONFI_DEVICE_NO_OF_LUNS__ONFI_DEVICE) { /* ONFI 1.0 NAND */
+               if (FAIL == get_onfi_nand_para())
+                       return FAIL;
+       } else if (DeviceInfo.wDeviceMaker == 0xEC) { /* Samsung NAND */
+               get_samsung_nand_para();
+       } else if (DeviceInfo.wDeviceMaker == 0x98) { /* Toshiba NAND */
+               get_toshiba_nand_para();
+       } else if (DeviceInfo.wDeviceMaker == 0xAD) { /* Hynix NAND */
+               get_hynix_nand_para();
+       } else {
+               DeviceInfo.wTotalBlocks = GLOB_HWCTL_DEFAULT_BLKS;
+       }
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Dump timing register values:"
+                       "acc_clks: %d, re_2_we: %d, we_2_re: %d,"
+                       "addr_2_data: %d, rdwr_en_lo_cnt: %d, "
+                       "rdwr_en_hi_cnt: %d, cs_setup_cnt: %d\n",
+                       ioread32(FlashReg + ACC_CLKS),
+                       ioread32(FlashReg + RE_2_WE),
+                       ioread32(FlashReg + WE_2_RE),
+                       ioread32(FlashReg + ADDR_2_DATA),
+                       ioread32(FlashReg + RDWR_EN_LO_CNT),
+                       ioread32(FlashReg + RDWR_EN_HI_CNT),
+                       ioread32(FlashReg + CS_SETUP_CNT));
+
+       DeviceInfo.wHWRevision = ioread32(FlashReg + REVISION);
+       DeviceInfo.wHWFeatures = ioread32(FlashReg + FEATURES);
+
+       DeviceInfo.wDeviceMainAreaSize =
+               ioread32(FlashReg + DEVICE_MAIN_AREA_SIZE);
+       DeviceInfo.wDeviceSpareAreaSize =
+               ioread32(FlashReg + DEVICE_SPARE_AREA_SIZE);
+
+       DeviceInfo.wPageDataSize =
+               ioread32(FlashReg + LOGICAL_PAGE_DATA_SIZE);
+
+       /* Note: When using the Micon 4K NAND device, the controller will report
+        * Page Spare Size as 216 bytes. But Micron's Spec say it's 218 bytes.
+        * And if force set it to 218 bytes, the controller can not work
+        * correctly. So just let it be. But keep in mind that this bug may
+        * cause
+        * other problems in future.       - Yunpeng  2008-10-10
+        */
+       DeviceInfo.wPageSpareSize =
+               ioread32(FlashReg + LOGICAL_PAGE_SPARE_SIZE);
+
+       DeviceInfo.wPagesPerBlock = ioread32(FlashReg + PAGES_PER_BLOCK);
+
+       DeviceInfo.wPageSize =
+           DeviceInfo.wPageDataSize + DeviceInfo.wPageSpareSize;
+       DeviceInfo.wBlockSize =
+           DeviceInfo.wPageSize * DeviceInfo.wPagesPerBlock;
+       DeviceInfo.wBlockDataSize =
+           DeviceInfo.wPagesPerBlock * DeviceInfo.wPageDataSize;
+
+       DeviceInfo.wDeviceWidth = ioread32(FlashReg + DEVICE_WIDTH);
+       DeviceInfo.wDeviceType =
+               ((ioread32(FlashReg + DEVICE_WIDTH) > 0) ? 16 : 8);
+
+       DeviceInfo.wDevicesConnected = ioread32(FlashReg + DEVICES_CONNECTED);
+
+       DeviceInfo.wSpareSkipBytes =
+               ioread32(FlashReg + SPARE_AREA_SKIP_BYTES) *
+               DeviceInfo.wDevicesConnected;
+
+       DeviceInfo.nBitsInPageNumber =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wPagesPerBlock);
+       DeviceInfo.nBitsInPageDataSize =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wPageDataSize);
+       DeviceInfo.nBitsInBlockDataSize =
+               (u8)GLOB_Calc_Used_Bits(DeviceInfo.wBlockDataSize);
+
+       set_ecc_config();
+
+       no_of_planes = ioread32(FlashReg + NUMBER_OF_PLANES) &
+               NUMBER_OF_PLANES__VALUE;
+
+       switch (no_of_planes) {
+       case 0:
+       case 1:
+       case 3:
+       case 7:
+               DeviceInfo.bPlaneNum = no_of_planes + 1;
+               break;
+       default:
+               status = FAIL;
+               break;
+       }
+
+       find_valid_banks();
+
+       detect_partition_feature();
+
+       dump_device_info();
+
+       return status;
+}
+
+u16 NAND_UnlockArrayAll(void)
+{
+       u64 start_addr, end_addr;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       start_addr = 0;
+       end_addr = ((u64)DeviceInfo.wBlockSize *
+               (DeviceInfo.wTotalBlocks - 1)) >>
+               DeviceInfo.nBitsInPageDataSize;
+
+       index_addr((u32)(MODE_10 | (u32)start_addr), 0x10);
+       index_addr((u32)(MODE_10 | (u32)end_addr), 0x11);
+
+       return PASS;
+}
+
+void NAND_LLD_Enable_Disable_Interrupts(u16 INT_ENABLE)
+{
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (INT_ENABLE)
+               iowrite32(1, FlashReg + GLOBAL_INT_ENABLE);
+       else
+               iowrite32(0, FlashReg + GLOBAL_INT_ENABLE);
+}
+
+u16 NAND_Erase_Block(u32 block)
+{
+       u16 status = PASS;
+       u64 flash_add;
+       u16 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (block >= DeviceInfo.wTotalBlocks)
+               status = FAIL;
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+
+               iowrite32(INTR_STATUS0__ERASE_COMP | INTR_STATUS0__ERASE_FAIL,
+                       FlashReg + intr_status);
+
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)), 1);
+
+               while (!(ioread32(FlashReg + intr_status) &
+                       (INTR_STATUS0__ERASE_COMP | INTR_STATUS0__ERASE_FAIL)))
+                       ;
+
+               if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__ERASE_FAIL)
+                       status = FAIL;
+
+               iowrite32(INTR_STATUS0__ERASE_COMP | INTR_STATUS0__ERASE_FAIL,
+                       FlashReg + intr_status);
+       }
+
+       return status;
+}
+
+static u32 Boundary_Check_Block_Page(u32 block, u16 page,
+                                               u16 page_count)
+{
+       u32 status = PASS;
+
+       if (block >= DeviceInfo.wTotalBlocks)
+               status = FAIL;
+
+       if (page + page_count > DeviceInfo.wPagesPerBlock)
+               status = FAIL;
+
+       return status;
+}
+
+u16 NAND_Read_Page_Spare(u8 *read_data, u32 block, u16 page,
+                           u16 page_count)
+{
+       u32 status = PASS;
+       u32 i;
+       u64 flash_add;
+       u32 PageSpareSize = DeviceInfo.wPageSpareSize;
+       u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u8 *page_spare = buf_read_page_spare;
+
+       if (block >= DeviceInfo.wTotalBlocks) {
+               printk(KERN_ERR "block too big: %d\n", (int)block);
+               status = FAIL;
+       }
+
+       if (page >= DeviceInfo.wPagesPerBlock) {
+               printk(KERN_ERR "page too big: %d\n", page);
+               status = FAIL;
+       }
+
+       if (page_count > 1) {
+               printk(KERN_ERR "page count too big: %d\n", page_count);
+               status = FAIL;
+       }
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+               iowrite32(ioread32(FlashReg + intr_status),
+                       FlashReg + intr_status);
+
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+                       0x41);
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+                       0x2000 | page_count);
+               while (!(ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__LOAD_COMP))
+                       ;
+
+               iowrite32((u32)(MODE_01 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+                       FlashMem);
+
+               for (i = 0; i < (PageSpareSize / 4); i++)
+                       *((u32 *)page_spare + i) =
+                                       ioread32(FlashMem + 0x10);
+
+               if (enable_ecc) {
+                       for (i = 0; i < spareFlagBytes; i++)
+                               read_data[i] =
+                                       page_spare[PageSpareSize -
+                                               spareFlagBytes + i];
+                       for (i = 0; i < (PageSpareSize - spareFlagBytes); i++)
+                               read_data[spareFlagBytes + i] =
+                                                       page_spare[i];
+               } else {
+                       for (i = 0; i < PageSpareSize; i++)
+                               read_data[i] = page_spare[i];
+               }
+
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+       }
+
+       return status;
+}
+
+/* No use function. Should be removed later */
+u16 NAND_Write_Page_Spare(u8 *write_data, u32 block, u16 page,
+                            u16 page_count)
+{
+       printk(KERN_ERR
+              "Error! This function (NAND_Write_Page_Spare) should never"
+               " be called!\n");
+       return ERR;
+}
+
+/* op value:  0 - DDMA read;  1 - DDMA write */
+static void ddma_trans(u8 *data, u64 flash_add,
+                       u32 flash_bank, int op, u32 numPages)
+{
+       u32 data_addr;
+
+       /* Map virtual address to bus address for DDMA */
+       data_addr = virt_to_bus(data);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+               (u16)(2 << 12) | (op << 8) | numPages);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               ((u16)(0x0FFFF & (data_addr >> 16)) << 8)),
+               (u16)(2 << 12) | (2 << 8) | 0);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               ((u16)(0x0FFFF & data_addr) << 8)),
+               (u16)(2 << 12) | (3 << 8) | 0);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (1 << 16) | (0x40 << 8)),
+               (u16)(2 << 12) | (4 << 8) | 0);
+}
+
+/* If data in buf are all 0xff, then return 1; otherwise return 0 */
+static int check_all_1(u8 *buf)
+{
+       int i, j, cnt;
+
+       for (i = 0; i < DeviceInfo.wPageDataSize; i++) {
+               if (buf[i] != 0xff) {
+                       cnt = 0;
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "the first non-0xff data byte is: %d\n", i);
+                       for (j = i; j < DeviceInfo.wPageDataSize; j++) {
+                               nand_dbg_print(NAND_DBG_WARN, "0x%x ", buf[j]);
+                               cnt++;
+                               if (cnt > 8)
+                                       break;
+                       }
+                       nand_dbg_print(NAND_DBG_WARN, "\n");
+                       return 0;
+               }
+       }
+
+       return 1;
+}
+
+static int do_ecc_new(unsigned long bank, u8 *buf,
+                               u32 block, u16 page)
+{
+       int status = PASS;
+       u16 err_page = 0;
+       u16 err_byte;
+       u8 err_sect;
+       u8 err_dev;
+       u16 err_fix_info;
+       u16 err_addr;
+       u32 ecc_sect_size;
+       u8 *err_pos;
+       u32 err_page_addr[4] = {ERR_PAGE_ADDR0,
+               ERR_PAGE_ADDR1, ERR_PAGE_ADDR2, ERR_PAGE_ADDR3};
+
+       ecc_sect_size = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+
+       do {
+               err_page = ioread32(FlashReg + err_page_addr[bank]);
+               err_addr = ioread32(FlashReg + ECC_ERROR_ADDRESS);
+               err_byte = err_addr & ECC_ERROR_ADDRESS__OFFSET;
+               err_sect = ((err_addr & ECC_ERROR_ADDRESS__SECTOR_NR) >> 12);
+               err_fix_info = ioread32(FlashReg + ERR_CORRECTION_INFO);
+               err_dev = ((err_fix_info & ERR_CORRECTION_INFO__DEVICE_NR)
+                       >> 8);
+               if (err_fix_info & ERR_CORRECTION_INFO__ERROR_TYPE) {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "%s, Line %d Uncorrectable ECC error "
+                               "when read block %d page %d."
+                               "PTN_INTR register: 0x%x "
+                               "err_page: %d, err_sect: %d, err_byte: %d, "
+                               "err_dev: %d, ecc_sect_size: %d, "
+                               "err_fix_info: 0x%x\n",
+                               __FILE__, __LINE__, block, page,
+                               ioread32(FlashReg + PTN_INTR),
+                               err_page, err_sect, err_byte, err_dev,
+                               ecc_sect_size, (u32)err_fix_info);
+
+                       if (check_all_1(buf))
+                               nand_dbg_print(NAND_DBG_WARN, "%s, Line %d"
+                                              "All 0xff!\n",
+                                              __FILE__, __LINE__);
+                       else
+                               nand_dbg_print(NAND_DBG_WARN, "%s, Line %d"
+                                              "Not all 0xff!\n",
+                                              __FILE__, __LINE__);
+                       status = FAIL;
+               } else {
+                       nand_dbg_print(NAND_DBG_WARN,
+                               "%s, Line %d Found ECC error "
+                               "when read block %d page %d."
+                               "err_page: %d, err_sect: %d, err_byte: %d, "
+                               "err_dev: %d, ecc_sect_size: %d, "
+                               "err_fix_info: 0x%x\n",
+                               __FILE__, __LINE__, block, page,
+                               err_page, err_sect, err_byte, err_dev,
+                               ecc_sect_size, (u32)err_fix_info);
+                       if (err_byte < ECC_SECTOR_SIZE) {
+                               err_pos = buf +
+                                       (err_page - page) *
+                                       DeviceInfo.wPageDataSize +
+                                       err_sect * ecc_sect_size +
+                                       err_byte *
+                                       DeviceInfo.wDevicesConnected +
+                                       err_dev;
+
+                               *err_pos ^= err_fix_info &
+                                       ERR_CORRECTION_INFO__BYTEMASK;
+                       }
+               }
+       } while (!(err_fix_info & ERR_CORRECTION_INFO__LAST_ERR_INFO));
+
+       return status;
+}
+
+u16 NAND_Read_Page_Main_Polling(u8 *read_data,
+               u32 block, u16 page, u16 page_count)
+{
+       u32 status = PASS;
+       u64 flash_add;
+       u32 intr_status = 0;
+       u32 flash_bank;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u8 *read_data_l;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       intr_status = intr_status_addresses[flash_bank];
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       if (page_count > 1) {
+               read_data_l = read_data;
+               while (page_count > MAX_PAGES_PER_RW) {
+                       if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                               status = NAND_Multiplane_Read(read_data_l,
+                                       block, page, MAX_PAGES_PER_RW);
+                       else
+                               status = NAND_Pipeline_Read_Ahead_Polling(
+                                       read_data_l, block, page,
+                                       MAX_PAGES_PER_RW);
+
+                       if (status == FAIL)
+                               return status;
+
+                       read_data_l += DeviceInfo.wPageDataSize *
+                                       MAX_PAGES_PER_RW;
+                       page_count -= MAX_PAGES_PER_RW;
+                       page += MAX_PAGES_PER_RW;
+               }
+               if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                       status = NAND_Multiplane_Read(read_data_l,
+                                       block, page, page_count);
+               else
+                       status = NAND_Pipeline_Read_Ahead_Polling(
+                                       read_data_l, block, page, page_count);
+
+               return status;
+       }
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       ddma_trans(read_data, flash_add, flash_bank, 0, 1);
+
+       if (enable_ecc) {
+               while (!(ioread32(FlashReg + intr_status) &
+                       (INTR_STATUS0__ECC_TRANSACTION_DONE |
+                       INTR_STATUS0__ECC_ERR)))
+                       ;
+
+               if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__ECC_ERR) {
+                       iowrite32(INTR_STATUS0__ECC_ERR,
+                               FlashReg + intr_status);
+                       status = do_ecc_new(flash_bank, read_data,
+                                       block, page);
+               }
+
+               if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__ECC_TRANSACTION_DONE &
+                       INTR_STATUS0__ECC_ERR)
+                       iowrite32(INTR_STATUS0__ECC_TRANSACTION_DONE |
+                               INTR_STATUS0__ECC_ERR,
+                               FlashReg + intr_status);
+               else if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__ECC_TRANSACTION_DONE)
+                       iowrite32(INTR_STATUS0__ECC_TRANSACTION_DONE,
+                               FlashReg + intr_status);
+               else if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__ECC_ERR)
+                       iowrite32(INTR_STATUS0__ECC_ERR,
+                               FlashReg + intr_status);
+       } else {
+               while (!(ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__DMA_CMD_COMP))
+                       ;
+               iowrite32(INTR_STATUS0__DMA_CMD_COMP, FlashReg + intr_status);
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       return status;
+}
+
+u16 NAND_Pipeline_Read_Ahead_Polling(u8 *read_data,
+                       u32 block, u16 page, u16 page_count)
+{
+       u32 status = PASS;
+       u32 NumPages = page_count;
+       u64 flash_add;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u32 ecc_done_OR_dma_comp;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       if (page_count < 2)
+               status = FAIL;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               *DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+               iowrite32(ioread32(FlashReg + intr_status),
+                       FlashReg + intr_status);
+
+               iowrite32(1, FlashReg + DMA_ENABLE);
+               while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+                       ;
+
+               iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+               ddma_trans(read_data, flash_add, flash_bank, 0, NumPages);
+
+               ecc_done_OR_dma_comp = 0;
+               while (1) {
+                       if (enable_ecc) {
+                               while (!ioread32(FlashReg + intr_status))
+                                       ;
+
+                               if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_ERR) {
+                                       iowrite32(INTR_STATUS0__ECC_ERR,
+                                               FlashReg + intr_status);
+                                       status = do_ecc_new(flash_bank,
+                                               read_data, block, page);
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__DMA_CMD_COMP) {
+                                       iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                               FlashReg + intr_status);
+
+                                       if (1 == ecc_done_OR_dma_comp)
+                                               break;
+
+                                       ecc_done_OR_dma_comp = 1;
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE) {
+                                       iowrite32(
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE,
+                                       FlashReg + intr_status);
+
+                                       if (1 == ecc_done_OR_dma_comp)
+                                               break;
+
+                                       ecc_done_OR_dma_comp = 1;
+                               }
+                       } else {
+                               while (!(ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__DMA_CMD_COMP))
+                                       ;
+
+                               iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                       FlashReg + intr_status);
+                               break;
+                       }
+
+                       iowrite32((~INTR_STATUS0__ECC_ERR) &
+                               (~INTR_STATUS0__ECC_TRANSACTION_DONE) &
+                               (~INTR_STATUS0__DMA_CMD_COMP),
+                               FlashReg + intr_status);
+
+               }
+
+               iowrite32(ioread32(FlashReg + intr_status),
+                       FlashReg + intr_status);
+
+               iowrite32(0, FlashReg + DMA_ENABLE);
+
+               while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+                       ;
+       }
+       return status;
+}
+
+u16 NAND_Read_Page_Main(u8 *read_data, u32 block, u16 page,
+                          u16 page_count)
+{
+       u32 status = PASS;
+       u64 flash_add;
+       u32 intr_status = 0;
+       u32 flash_bank;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       int ret;
+       u8 *read_data_l;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       intr_status = intr_status_addresses[flash_bank];
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       if (page_count > 1) {
+               read_data_l = read_data;
+               while (page_count > MAX_PAGES_PER_RW) {
+                       if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                               status = NAND_Multiplane_Read(read_data_l,
+                                       block, page, MAX_PAGES_PER_RW);
+                       else
+                               status = NAND_Pipeline_Read_Ahead(
+                                       read_data_l, block, page,
+                                       MAX_PAGES_PER_RW);
+
+                       if (status == FAIL)
+                               return status;
+
+                       read_data_l += DeviceInfo.wPageDataSize *
+                                       MAX_PAGES_PER_RW;
+                       page_count -= MAX_PAGES_PER_RW;
+                       page += MAX_PAGES_PER_RW;
+               }
+               if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                       status = NAND_Multiplane_Read(read_data_l,
+                                       block, page, page_count);
+               else
+                       status = NAND_Pipeline_Read_Ahead(
+                                       read_data_l, block, page, page_count);
+
+               return status;
+       }
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       /* Fill the mrst_nand_info structure */
+       info.state = INT_READ_PAGE_MAIN;
+       info.read_data = read_data;
+       info.flash_bank = flash_bank;
+       info.block = block;
+       info.page = page;
+       info.ret = PASS;
+
+       ddma_trans(read_data, flash_add, flash_bank, 0, 1);
+
+       iowrite32(1, FlashReg + GLOBAL_INT_ENABLE); /* Enable Interrupt */
+
+       ret = wait_for_completion_timeout(&info.complete, 10 * HZ);
+       if (!ret) {
+               printk(KERN_ERR "Wait for completion timeout "
+                       "in %s, Line %d\n", __FILE__, __LINE__);
+               status = ERR;
+       } else {
+               status = info.ret;
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       return status;
+}
+
+void Conv_Spare_Data_Log2Phy_Format(u8 *data)
+{
+       int i;
+       const u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       const u32 PageSpareSize  = DeviceInfo.wPageSpareSize;
+
+       if (enable_ecc) {
+               for (i = spareFlagBytes - 1; i >= 0; i++)
+                       data[PageSpareSize - spareFlagBytes + i] = data[i];
+       }
+}
+
+void Conv_Spare_Data_Phy2Log_Format(u8 *data)
+{
+       int i;
+       const u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       const u32 PageSpareSize = DeviceInfo.wPageSpareSize;
+
+       if (enable_ecc) {
+               for (i = 0; i < spareFlagBytes; i++)
+                       data[i] = data[PageSpareSize - spareFlagBytes + i];
+       }
+}
+
+
+void Conv_Main_Spare_Data_Log2Phy_Format(u8 *data, u16 page_count)
+{
+       const u32 PageSize = DeviceInfo.wPageSize;
+       const u32 PageDataSize = DeviceInfo.wPageDataSize;
+       const u32 eccBytes = DeviceInfo.wECCBytesPerSector;
+       const u32 spareSkipBytes = DeviceInfo.wSpareSkipBytes;
+       const u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       u32 eccSectorSize;
+       u32 page_offset;
+       int i, j;
+
+       eccSectorSize = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+       if (enable_ecc) {
+               while (page_count > 0) {
+                       page_offset = (page_count - 1) * PageSize;
+                       j = (DeviceInfo.wPageDataSize / eccSectorSize);
+                       for (i = spareFlagBytes - 1; i >= 0; i--)
+                               data[page_offset +
+                                       (eccSectorSize + eccBytes) * j + i] =
+                                       data[page_offset + PageDataSize + i];
+                       for (j--; j >= 1; j--) {
+                               for (i = eccSectorSize - 1; i >= 0; i--)
+                                       data[page_offset +
+                                       (eccSectorSize + eccBytes) * j + i] =
+                                               data[page_offset +
+                                               eccSectorSize * j + i];
+                       }
+                       for (i = (PageSize - spareSkipBytes) - 1;
+                               i >= PageDataSize; i--)
+                               data[page_offset + i + spareSkipBytes] =
+                                       data[page_offset + i];
+                       page_count--;
+               }
+       }
+}
+
+void Conv_Main_Spare_Data_Phy2Log_Format(u8 *data, u16 page_count)
+{
+       const u32 PageSize = DeviceInfo.wPageSize;
+       const u32 PageDataSize = DeviceInfo.wPageDataSize;
+       const u32 eccBytes = DeviceInfo.wECCBytesPerSector;
+       const u32 spareSkipBytes = DeviceInfo.wSpareSkipBytes;
+       const u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       u32 eccSectorSize;
+       u32 page_offset;
+       int i, j;
+
+       eccSectorSize = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+       if (enable_ecc) {
+               while (page_count > 0) {
+                       page_offset = (page_count - 1) * PageSize;
+                       for (i = PageDataSize;
+                               i < PageSize - spareSkipBytes;
+                               i++)
+                               data[page_offset + i] =
+                                       data[page_offset + i +
+                                       spareSkipBytes];
+                       for (j = 1;
+                       j < DeviceInfo.wPageDataSize / eccSectorSize;
+                       j++) {
+                               for (i = 0; i < eccSectorSize; i++)
+                                       data[page_offset +
+                                       eccSectorSize * j + i] =
+                                               data[page_offset +
+                                               (eccSectorSize + eccBytes) * j
+                                               + i];
+                       }
+                       for (i = 0; i < spareFlagBytes; i++)
+                               data[page_offset + PageDataSize + i] =
+                                       data[page_offset +
+                                       (eccSectorSize + eccBytes) * j + i];
+                       page_count--;
+               }
+       }
+}
+
+/* Un-tested function */
+u16 NAND_Multiplane_Read(u8 *read_data, u32 block, u16 page,
+                           u16 page_count)
+{
+       u32 status = PASS;
+       u32 NumPages = page_count;
+       u64 flash_add;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u32 ecc_done_OR_dma_comp;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+               iowrite32(ioread32(FlashReg + intr_status),
+                       FlashReg + intr_status);
+
+               iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+               iowrite32(0x01, FlashReg + MULTIPLANE_OPERATION);
+
+               iowrite32(1, FlashReg + DMA_ENABLE);
+               while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+                       ;
+               index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                       (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+               ddma_trans(read_data, flash_add, flash_bank, 0, NumPages);
+
+               ecc_done_OR_dma_comp = 0;
+               while (1) {
+                       if (enable_ecc) {
+                               while (!ioread32(FlashReg + intr_status))
+                                       ;
+
+                               if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_ERR) {
+                                       iowrite32(INTR_STATUS0__ECC_ERR,
+                                               FlashReg + intr_status);
+                                       status = do_ecc_new(flash_bank,
+                                               read_data, block, page);
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__DMA_CMD_COMP) {
+                                       iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                               FlashReg + intr_status);
+
+                                       if (1 == ecc_done_OR_dma_comp)
+                                               break;
+
+                                       ecc_done_OR_dma_comp = 1;
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE) {
+                                       iowrite32(
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE,
+                                       FlashReg + intr_status);
+
+                                       if (1 == ecc_done_OR_dma_comp)
+                                               break;
+
+                                       ecc_done_OR_dma_comp = 1;
+                               }
+                       } else {
+                               while (!(ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__DMA_CMD_COMP))
+                                       ;
+                               iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                       FlashReg + intr_status);
+                               break;
+                       }
+
+                       iowrite32((~INTR_STATUS0__ECC_ERR) &
+                               (~INTR_STATUS0__ECC_TRANSACTION_DONE) &
+                               (~INTR_STATUS0__DMA_CMD_COMP),
+                               FlashReg + intr_status);
+
+               }
+
+               iowrite32(ioread32(FlashReg + intr_status),
+                       FlashReg + intr_status);
+
+               iowrite32(0, FlashReg + DMA_ENABLE);
+
+               while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+                       ;
+
+               iowrite32(0, FlashReg + MULTIPLANE_OPERATION);
+       }
+
+       return status;
+}
+
+u16 NAND_Pipeline_Read_Ahead(u8 *read_data, u32 block,
+                               u16 page, u16 page_count)
+{
+       u32 status = PASS;
+       u32 NumPages = page_count;
+       u64 flash_add;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       int ret;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       if (page_count < 2)
+               status = FAIL;
+
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               *DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       intr_status = intr_status_addresses[flash_bank];
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       /* Fill the mrst_nand_info structure */
+       info.state = INT_PIPELINE_READ_AHEAD;
+       info.read_data = read_data;
+       info.flash_bank = flash_bank;
+       info.block = block;
+       info.page = page;
+       info.ret = PASS;
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+
+       ddma_trans(read_data, flash_add, flash_bank, 0, NumPages);
+
+       iowrite32(1, FlashReg + GLOBAL_INT_ENABLE); /* Enable Interrupt */
+
+       ret = wait_for_completion_timeout(&info.complete, 10 * HZ);
+       if (!ret) {
+               printk(KERN_ERR "Wait for completion timeout "
+                       "in %s, Line %d\n", __FILE__, __LINE__);
+               status = ERR;
+       } else {
+               status = info.ret;
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       return status;
+}
+
+
+u16 NAND_Write_Page_Main(u8 *write_data, u32 block, u16 page,
+                           u16 page_count)
+{
+       u32 status = PASS;
+       u64 flash_add;
+       u32 intr_status = 0;
+       u32 flash_bank;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       int ret;
+       u8 *write_data_l;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       intr_status = intr_status_addresses[flash_bank];
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       iowrite32(INTR_STATUS0__PROGRAM_COMP |
+               INTR_STATUS0__PROGRAM_FAIL, FlashReg + intr_status);
+
+       if (page_count > 1) {
+               write_data_l = write_data;
+               while (page_count > MAX_PAGES_PER_RW) {
+                       if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                               status = NAND_Multiplane_Write(write_data_l,
+                                       block, page, MAX_PAGES_PER_RW);
+                       else
+                               status = NAND_Pipeline_Write_Ahead(
+                                       write_data_l, block, page,
+                                       MAX_PAGES_PER_RW);
+                       if (status == FAIL)
+                               return status;
+
+                       write_data_l += DeviceInfo.wPageDataSize *
+                                       MAX_PAGES_PER_RW;
+                       page_count -= MAX_PAGES_PER_RW;
+                       page += MAX_PAGES_PER_RW;
+               }
+               if (ioread32(FlashReg + MULTIPLANE_OPERATION))
+                       status = NAND_Multiplane_Write(write_data_l,
+                               block, page, page_count);
+               else
+                       status = NAND_Pipeline_Write_Ahead(write_data_l,
+                               block, page, page_count);
+
+               return status;
+       }
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       /* Fill the mrst_nand_info structure */
+       info.state = INT_WRITE_PAGE_MAIN;
+       info.write_data = write_data;
+       info.flash_bank = flash_bank;
+       info.block = block;
+       info.page = page;
+       info.ret = PASS;
+
+       ddma_trans(write_data, flash_add, flash_bank, 1, 1);
+
+       iowrite32(1, FlashReg + GLOBAL_INT_ENABLE); /* Enable interrupt */
+
+       ret = wait_for_completion_timeout(&info.complete, 10 * HZ);
+       if (!ret) {
+               printk(KERN_ERR "Wait for completion timeout "
+                       "in %s, Line %d\n", __FILE__, __LINE__);
+               status = ERR;
+       } else {
+               status = info.ret;
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+       while (ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG)
+               ;
+
+       return status;
+}
+
+void NAND_ECC_Ctrl(int enable)
+{
+       if (enable) {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Will enable ECC in %s, Line %d, Function: %s\n",
+                       __FILE__, __LINE__, __func__);
+               iowrite32(1, FlashReg + ECC_ENABLE);
+               enable_ecc = 1;
+       } else {
+               nand_dbg_print(NAND_DBG_WARN,
+                       "Will disable ECC in %s, Line %d, Function: %s\n",
+                       __FILE__, __LINE__, __func__);
+               iowrite32(0, FlashReg + ECC_ENABLE);
+               enable_ecc = 0;
+       }
+}
+
+u16 NAND_Write_Page_Main_Spare(u8 *write_data, u32 block,
+                                       u16 page, u16 page_count)
+{
+       u32 status = PASS;
+       u32 i, j, page_num = 0;
+       u32 PageSize = DeviceInfo.wPageSize;
+       u32 PageDataSize = DeviceInfo.wPageDataSize;
+       u32 eccBytes = DeviceInfo.wECCBytesPerSector;
+       u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       u32 spareSkipBytes  = DeviceInfo.wSpareSkipBytes;
+       u64 flash_add;
+       u32 eccSectorSize;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u8 *page_main_spare = buf_write_page_main_spare;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       eccSectorSize = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+
+               iowrite32(1, FlashReg + TRANSFER_SPARE_REG);
+
+               while ((status != FAIL) && (page_count > 0)) {
+                       flash_add = (u64)(block %
+                       (DeviceInfo.wTotalBlocks / totalUsedBanks)) *
+                       DeviceInfo.wBlockDataSize +
+                       (u64)page * DeviceInfo.wPageDataSize;
+
+                       iowrite32(ioread32(FlashReg + intr_status),
+                               FlashReg + intr_status);
+
+                       iowrite32((u32)(MODE_01 | (flash_bank << 24) |
+                               (flash_add >>
+                               DeviceInfo.nBitsInPageDataSize)),
+                               FlashMem);
+
+                       if (enable_ecc) {
+                               for (j = 0;
+                                    j <
+                                    DeviceInfo.wPageDataSize / eccSectorSize;
+                                    j++) {
+                                       for (i = 0; i < eccSectorSize; i++)
+                                               page_main_spare[(eccSectorSize +
+                                                                eccBytes) * j +
+                                                               i] =
+                                                   write_data[eccSectorSize *
+                                                              j + i];
+
+                                       for (i = 0; i < eccBytes; i++)
+                                               page_main_spare[(eccSectorSize +
+                                                                eccBytes) * j +
+                                                               eccSectorSize +
+                                                               i] =
+                                                   write_data[PageDataSize +
+                                                              spareFlagBytes +
+                                                              eccBytes * j +
+                                                              i];
+                               }
+
+                               for (i = 0; i < spareFlagBytes; i++)
+                                       page_main_spare[(eccSectorSize +
+                                                        eccBytes) * j + i] =
+                                           write_data[PageDataSize + i];
+
+                               for (i = PageSize - 1; i >= PageDataSize +
+                                                       spareSkipBytes; i--)
+                                       page_main_spare[i] = page_main_spare[i -
+                                                               spareSkipBytes];
+
+                               for (i = PageDataSize; i < PageDataSize +
+                                                       spareSkipBytes; i++)
+                                       page_main_spare[i] = 0xff;
+
+                               for (i = 0; i < PageSize / 4; i++)
+                                       iowrite32(
+                                       *((u32 *)page_main_spare + i),
+                                       FlashMem + 0x10);
+                       } else {
+
+                               for (i = 0; i < PageSize / 4; i++)
+                                       iowrite32(*((u32 *)write_data + i),
+                                               FlashMem + 0x10);
+                       }
+
+                       while (!(ioread32(FlashReg + intr_status) &
+                               (INTR_STATUS0__PROGRAM_COMP |
+                               INTR_STATUS0__PROGRAM_FAIL)))
+                               ;
+
+                       if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__PROGRAM_FAIL)
+                               status = FAIL;
+
+                       iowrite32(ioread32(FlashReg + intr_status),
+                                       FlashReg + intr_status);
+
+                       page_num++;
+                       page_count--;
+                       write_data += PageSize;
+               }
+
+               iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+       }
+
+       return status;
+}
+
+u16 NAND_Read_Page_Main_Spare(u8 *read_data, u32 block, u16 page,
+                                u16 page_count)
+{
+       u32 status = PASS;
+       u32 i, j;
+       u64 flash_add = 0;
+       u32 PageSize = DeviceInfo.wPageSize;
+       u32 PageDataSize = DeviceInfo.wPageDataSize;
+       u32 PageSpareSize = DeviceInfo.wPageSpareSize;
+       u32 eccBytes = DeviceInfo.wECCBytesPerSector;
+       u32 spareFlagBytes = DeviceInfo.wNumPageSpareFlag;
+       u32 spareSkipBytes  = DeviceInfo.wSpareSkipBytes;
+       u32 eccSectorSize;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u8 *read_data_l = read_data;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u8 *page_main_spare = buf_read_page_main_spare;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       eccSectorSize = ECC_SECTOR_SIZE * (DeviceInfo.wDevicesConnected);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       if (status == PASS) {
+               intr_status = intr_status_addresses[flash_bank];
+
+               iowrite32(1, FlashReg + TRANSFER_SPARE_REG);
+
+               iowrite32(ioread32(FlashReg + intr_status),
+                               FlashReg + intr_status);
+
+               while ((status != FAIL) && (page_count > 0)) {
+                       flash_add = (u64)(block %
+                               (DeviceInfo.wTotalBlocks / totalUsedBanks))
+                               * DeviceInfo.wBlockDataSize +
+                               (u64)page * DeviceInfo.wPageDataSize;
+
+                       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                               (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+                               0x43);
+                       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+                               (flash_add >> DeviceInfo.nBitsInPageDataSize)),
+                               0x2000 | page_count);
+
+                       while (!(ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__LOAD_COMP))
+                               ;
+
+                       iowrite32((u32)(MODE_01 | (flash_bank << 24) |
+                               (flash_add >>
+                               DeviceInfo.nBitsInPageDataSize)),
+                               FlashMem);
+
+                       for (i = 0; i < PageSize / 4; i++)
+                               *(((u32 *)page_main_spare) + i) =
+                                       ioread32(FlashMem + 0x10);
+
+                       if (enable_ecc) {
+                               for (i = PageDataSize;  i < PageSize -
+                                                       spareSkipBytes; i++)
+                                       page_main_spare[i] = page_main_spare[i +
+                                                               spareSkipBytes];
+
+                               for (j = 0;
+                               j < DeviceInfo.wPageDataSize / eccSectorSize;
+                               j++) {
+
+                                       for (i = 0; i < eccSectorSize; i++)
+                                               read_data_l[eccSectorSize * j +
+                                                           i] =
+                                                   page_main_spare[
+                                                       (eccSectorSize +
+                                                       eccBytes) * j + i];
+
+                                       for (i = 0; i < eccBytes; i++)
+                                               read_data_l[PageDataSize +
+                                                           spareFlagBytes +
+                                                           eccBytes * j + i] =
+                                                   page_main_spare[
+                                                       (eccSectorSize +
+                                                       eccBytes) * j +
+                                                       eccSectorSize + i];
+                               }
+
+                               for (i = 0; i < spareFlagBytes; i++)
+                                       read_data_l[PageDataSize + i] =
+                                           page_main_spare[(eccSectorSize +
+                                                            eccBytes) * j + i];
+                       } else {
+                               for (i = 0; i < (PageDataSize + PageSpareSize);
+                                    i++)
+                                       read_data_l[i] = page_main_spare[i];
+
+                       }
+
+                       if (enable_ecc) {
+                               while (!(ioread32(FlashReg + intr_status) &
+                                       (INTR_STATUS0__ECC_TRANSACTION_DONE |
+                                       INTR_STATUS0__ECC_ERR)))
+                                       ;
+
+                               if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_ERR) {
+                                       iowrite32(INTR_STATUS0__ECC_ERR,
+                                               FlashReg + intr_status);
+                                       status = do_ecc_new(flash_bank,
+                                               read_data, block, page);
+                               }
+
+                               if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE &
+                                       INTR_STATUS0__ECC_ERR) {
+                                       iowrite32(INTR_STATUS0__ECC_ERR |
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE,
+                                       FlashReg + intr_status);
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE) {
+                                       iowrite32(
+                                       INTR_STATUS0__ECC_TRANSACTION_DONE,
+                                       FlashReg + intr_status);
+                               } else if (ioread32(FlashReg + intr_status) &
+                                       INTR_STATUS0__ECC_ERR) {
+                                       iowrite32(INTR_STATUS0__ECC_ERR,
+                                               FlashReg + intr_status);
+                               }
+                       }
+
+                       page++;
+                       page_count--;
+                       read_data_l += PageSize;
+               }
+       }
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+
+       return status;
+}
+
+u16 NAND_Pipeline_Write_Ahead(u8 *write_data, u32 block,
+                       u16 page, u16 page_count)
+{
+       u16 status = PASS;
+       u32 NumPages = page_count;
+       u64 flash_add;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       int ret;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+
+       if (page_count < 2)
+               status = FAIL;
+
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       intr_status = intr_status_addresses[flash_bank];
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       /* Fill the mrst_nand_info structure */
+       info.state = INT_PIPELINE_WRITE_AHEAD;
+       info.write_data = write_data;
+       info.flash_bank = flash_bank;
+       info.block = block;
+       info.page = page;
+       info.ret = PASS;
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+
+       ddma_trans(write_data, flash_add, flash_bank, 1, NumPages);
+
+       iowrite32(1, FlashReg + GLOBAL_INT_ENABLE); /* Enable interrupt */
+
+       ret = wait_for_completion_timeout(&info.complete, 10 * HZ);
+       if (!ret) {
+               printk(KERN_ERR "Wait for completion timeout "
+                       "in %s, Line %d\n", __FILE__, __LINE__);
+               status = ERR;
+       } else {
+               status = info.ret;
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       return status;
+}
+
+/* Un-tested function */
+u16 NAND_Multiplane_Write(u8 *write_data, u32 block, u16 page,
+                            u16 page_count)
+{
+       u16 status = PASS;
+       u32 NumPages = page_count;
+       u64 flash_add;
+       u32 flash_bank;
+       u32 intr_status = 0;
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u16 status2 = PASS;
+       u32 t;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       status = Boundary_Check_Block_Page(block, page, page_count);
+       if (status != PASS)
+               return status;
+
+       flash_add = (u64)(block % (DeviceInfo.wTotalBlocks / totalUsedBanks))
+               * DeviceInfo.wBlockDataSize +
+               (u64)page * DeviceInfo.wPageDataSize;
+
+       flash_bank = block / (DeviceInfo.wTotalBlocks / totalUsedBanks);
+
+       intr_status = intr_status_addresses[flash_bank];
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+       iowrite32(0x01, FlashReg + MULTIPLANE_OPERATION);
+
+       iowrite32(1, FlashReg + DMA_ENABLE);
+       while (!(ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + TRANSFER_SPARE_REG);
+
+       index_addr((u32)(MODE_10 | (flash_bank << 24) |
+               (flash_add >> DeviceInfo.nBitsInPageDataSize)), 0x42);
+
+       ddma_trans(write_data, flash_add, flash_bank, 1, NumPages);
+
+       while (1) {
+               while (!ioread32(FlashReg + intr_status))
+                       ;
+
+               if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__DMA_CMD_COMP) {
+                       iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                               FlashReg + intr_status);
+                       status = PASS;
+                       if (status2 == FAIL)
+                               status = FAIL;
+                       break;
+               } else if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__PROGRAM_FAIL) {
+                       status2 = FAIL;
+                       status = FAIL;
+                       t = ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__PROGRAM_FAIL;
+                       iowrite32(t, FlashReg + intr_status);
+               } else {
+                       iowrite32((~INTR_STATUS0__PROGRAM_FAIL) &
+                               (~INTR_STATUS0__DMA_CMD_COMP),
+                               FlashReg + intr_status);
+               }
+       }
+
+       iowrite32(ioread32(FlashReg + intr_status), FlashReg + intr_status);
+
+       iowrite32(0, FlashReg + DMA_ENABLE);
+
+       while ((ioread32(FlashReg + DMA_ENABLE) & DMA_ENABLE__FLAG))
+               ;
+
+       iowrite32(0, FlashReg + MULTIPLANE_OPERATION);
+
+       return status;
+}
+
+
+#if CMD_DMA
+static irqreturn_t cdma_isr(int irq, void *dev_id)
+{
+       struct mrst_nand_info *dev = dev_id;
+       int first_failed_cmd;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       if (!is_cdma_interrupt())
+               return IRQ_NONE;
+
+       /* Disable controller interrupts */
+       iowrite32(0, FlashReg + GLOBAL_INT_ENABLE);
+       GLOB_FTL_Event_Status(&first_failed_cmd);
+       complete(&dev->complete);
+
+       return IRQ_HANDLED;
+}
+#else
+static void handle_nand_int_read(struct mrst_nand_info *dev)
+{
+       u32 intr_status_addresses[4] = {INTR_STATUS0,
+               INTR_STATUS1, INTR_STATUS2, INTR_STATUS3};
+       u32 intr_status;
+       u32 ecc_done_OR_dma_comp = 0;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       dev->ret = PASS;
+       intr_status = intr_status_addresses[dev->flash_bank];
+
+       while (1) {
+               if (enable_ecc) {
+                       if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__ECC_ERR) {
+                               iowrite32(INTR_STATUS0__ECC_ERR,
+                                       FlashReg + intr_status);
+                               dev->ret = do_ecc_new(dev->flash_bank,
+                                               dev->read_data,
+                                               dev->block, dev->page);
+                       } else if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__DMA_CMD_COMP) {
+                               iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                       FlashReg + intr_status);
+                               if (1 == ecc_done_OR_dma_comp)
+                                       break;
+                               ecc_done_OR_dma_comp = 1;
+                       } else if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__ECC_TRANSACTION_DONE) {
+                               iowrite32(INTR_STATUS0__ECC_TRANSACTION_DONE,
+                                       FlashReg + intr_status);
+                               if (1 == ecc_done_OR_dma_comp)
+                                       break;
+                               ecc_done_OR_dma_comp = 1;
+                       }
+               } else {
+                       if (ioread32(FlashReg + intr_status) &
+                               INTR_STATUS0__DMA_CMD_COMP) {
+                               iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                                       FlashReg + intr_status);
+                               break;
+                       } else {
+                               printk(KERN_ERR "Illegal INTS "
+                                       "(offset addr 0x%x) value: 0x%x\n",
+                                       intr_status,
+                                       ioread32(FlashReg + intr_status));
+                       }
+               }
+
+               iowrite32((~INTR_STATUS0__ECC_ERR) &
+               (~INTR_STATUS0__ECC_TRANSACTION_DONE) &
+               (~INTR_STATUS0__DMA_CMD_COMP),
+               FlashReg + intr_status);
+       }
+}
+
+static void handle_nand_int_write(struct mrst_nand_info *dev)
+{
+       u32 intr_status;
+       u32 intr[4] = {INTR_STATUS0, INTR_STATUS1,
+               INTR_STATUS2, INTR_STATUS3};
+       int status = PASS;
+
+       nand_dbg_print(NAND_DBG_DEBUG, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       dev->ret = PASS;
+       intr_status = intr[dev->flash_bank];
+
+       while (1) {
+               while (!ioread32(FlashReg + intr_status))
+                       ;
+
+               if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__DMA_CMD_COMP) {
+                       iowrite32(INTR_STATUS0__DMA_CMD_COMP,
+                               FlashReg + intr_status);
+                       if (FAIL == status)
+                               dev->ret = FAIL;
+                       break;
+               } else if (ioread32(FlashReg + intr_status) &
+                       INTR_STATUS0__PROGRAM_FAIL) {
+                       status = FAIL;
+                       iowrite32(INTR_STATUS0__PROGRAM_FAIL,
+                               FlashReg + intr_status);
+               } else {
+                       iowrite32((~INTR_STATUS0__PROGRAM_FAIL) &
+                               (~INTR_STATUS0__DMA_CMD_COMP),
+                               FlashReg + intr_status);
+               }
+       }
+}
+
+static irqreturn_t ddma_isr(int irq, void *dev_id)
+{
+       struct mrst_nand_info *dev = dev_id;
+       u32 int_mask, ints0, ints1, ints2, ints3, ints_offset;
+       u32 intr[4] = {INTR_STATUS0, INTR_STATUS1,
+               INTR_STATUS2, INTR_STATUS3};
+
+       int_mask = INTR_STATUS0__DMA_CMD_COMP |
+               INTR_STATUS0__ECC_TRANSACTION_DONE |
+               INTR_STATUS0__ECC_ERR |
+               INTR_STATUS0__PROGRAM_FAIL |
+               INTR_STATUS0__ERASE_FAIL;
+
+       ints0 = ioread32(FlashReg + INTR_STATUS0);
+       ints1 = ioread32(FlashReg + INTR_STATUS1);
+       ints2 = ioread32(FlashReg + INTR_STATUS2);
+       ints3 = ioread32(FlashReg + INTR_STATUS3);
+
+       ints_offset = intr[dev->flash_bank];
+
+       nand_dbg_print(NAND_DBG_DEBUG,
+               "INTR0: 0x%x, INTR1: 0x%x, INTR2: 0x%x, INTR3: 0x%x, "
+               "DMA_INTR: 0x%x, "
+               "dev->state: 0x%x, dev->flash_bank: %d\n",
+               ints0, ints1, ints2, ints3,
+               ioread32(FlashReg + DMA_INTR),
+               dev->state, dev->flash_bank);
+
+       if (!(ioread32(FlashReg + ints_offset) & int_mask)) {
+               iowrite32(ints0, FlashReg + INTR_STATUS0);
+               iowrite32(ints1, FlashReg + INTR_STATUS1);
+               iowrite32(ints2, FlashReg + INTR_STATUS2);
+               iowrite32(ints3, FlashReg + INTR_STATUS3);
+               nand_dbg_print(NAND_DBG_WARN,
+                       "ddma_isr: Invalid interrupt for NAND controller. "
+                       "Ignore it\n");
+               return IRQ_NONE;
+       }
+
+       switch (dev->state) {
+       case INT_READ_PAGE_MAIN:
+       case INT_PIPELINE_READ_AHEAD:
+               /* Disable controller interrupts */
+               iowrite32(0, FlashReg + GLOBAL_INT_ENABLE);
+               handle_nand_int_read(dev);
+               break;
+       case INT_WRITE_PAGE_MAIN:
+       case INT_PIPELINE_WRITE_AHEAD:
+               iowrite32(0, FlashReg + GLOBAL_INT_ENABLE);
+               handle_nand_int_write(dev);
+               break;
+       default:
+               printk(KERN_ERR "ddma_isr - Illegal state: 0x%x\n",
+                       dev->state);
+               return IRQ_NONE;
+       }
+
+       dev->state = INT_IDLE_STATE;
+       complete(&dev->complete);
+       return IRQ_HANDLED;
+}
+#endif
+
+static const struct pci_device_id nand_pci_ids[] = {
+       {
+        .vendor = 0x8086,
+        .device = 0x0809,
+        .subvendor = PCI_ANY_ID,
+        .subdevice = PCI_ANY_ID,
+        },
+       { /* end: all zeroes */ }
+};
+
+static int nand_pci_probe(struct pci_dev *dev, const struct pci_device_id *id)
+{
+       int ret = -ENODEV;
+       unsigned long csr_base;
+       unsigned long csr_len;
+       struct mrst_nand_info *pndev = &info;
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       ret = pci_enable_device(dev);
+       if (ret) {
+               printk(KERN_ERR "Spectra: pci_enable_device failed.\n");
+               return ret;
+       }
+
+       pci_set_master(dev);
+       pndev->dev = dev;
+
+       csr_base = pci_resource_start(dev, 0);
+       if (!csr_base) {
+               printk(KERN_ERR "Spectra: pci_resource_start failed!\n");
+               return -ENODEV;
+       }
+
+       csr_len = pci_resource_len(dev, 0);
+       if (!csr_len) {
+               printk(KERN_ERR "Spectra: pci_resource_len failed!\n");
+               return -ENODEV;
+       }
+
+       ret = pci_request_regions(dev, SPECTRA_NAND_NAME);
+       if (ret) {
+               printk(KERN_ERR "Spectra: Unable to request "
+                      "memory region\n");
+               goto failed_req_csr;
+       }
+
+       pndev->ioaddr = ioremap_nocache(csr_base, csr_len);
+       if (!pndev->ioaddr) {
+               printk(KERN_ERR "Spectra: Unable to remap memory region\n");
+               ret = -ENOMEM;
+               goto failed_remap_csr;
+       }
+       nand_dbg_print(NAND_DBG_DEBUG, "Spectra: CSR 0x%08lx -> 0x%p (0x%lx)\n",
+                      csr_base, pndev->ioaddr, csr_len);
+
+       init_completion(&pndev->complete);
+       nand_dbg_print(NAND_DBG_DEBUG, "Spectra: IRQ %d\n", dev->irq);
+
+#if CMD_DMA
+       if (request_irq(dev->irq, cdma_isr, IRQF_SHARED,
+                       SPECTRA_NAND_NAME, &info)) {
+               printk(KERN_ERR "Spectra: Unable to allocate IRQ\n");
+               ret = -ENODEV;
+               iounmap(pndev->ioaddr);
+               goto failed_remap_csr;
+       }
+#else
+       if (request_irq(dev->irq, ddma_isr, IRQF_SHARED,
+                       SPECTRA_NAND_NAME, &info)) {
+               printk(KERN_ERR "Spectra: Unable to allocate IRQ\n");
+               ret = -ENODEV;
+               iounmap(pndev->ioaddr);
+               goto failed_remap_csr;
+       }
+#endif
+
+       pci_set_drvdata(dev, pndev);
+
+       return 0;
+
+failed_remap_csr:
+       pci_release_regions(dev);
+failed_req_csr:
+
+       return ret;
+}
+
+static void nand_pci_remove(struct pci_dev *dev)
+{
+       struct mrst_nand_info *pndev = pci_get_drvdata(dev);
+
+       nand_dbg_print(NAND_DBG_WARN, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+#if CMD_DMA
+       free_irq(dev->irq, pndev);
+#endif
+       iounmap(pndev->ioaddr);
+       pci_release_regions(dev);
+       pci_disable_device(dev);
+}
+
+MODULE_DEVICE_TABLE(pci, nand_pci_ids);
+
+static struct pci_driver nand_pci_driver = {
+       .name = SPECTRA_NAND_NAME,
+       .id_table = nand_pci_ids,
+       .probe = nand_pci_probe,
+       .remove = nand_pci_remove,
+};
+
+int NAND_Flash_Init(void)
+{
+       int retval;
+       u32 int_mask;
+
+       nand_dbg_print(NAND_DBG_TRACE, "%s, Line %d, Function: %s\n",
+                      __FILE__, __LINE__, __func__);
+
+       FlashReg = ioremap_nocache(GLOB_HWCTL_REG_BASE,
+                       GLOB_HWCTL_REG_SIZE);
+       if (!FlashReg) {
+               printk(KERN_ERR "Spectra: ioremap_nocache failed!");
+               return -ENOMEM;
+       }
+       nand_dbg_print(NAND_DBG_WARN,
+               "Spectra: Remapped reg base address: "
+               "0x%p, len: %d\n",
+               FlashReg, GLOB_HWCTL_REG_SIZE);
+
+       FlashMem = ioremap_nocache(GLOB_HWCTL_MEM_BASE,
+                       GLOB_HWCTL_MEM_SIZE);
+       if (!FlashMem) {
+               printk(KERN_ERR "Spectra: ioremap_nocache failed!");
+               iounmap(FlashReg);
+               return -ENOMEM;
+       }
+       nand_dbg_print(NAND_DBG_WARN,
+               "Spectra: Remapped flash base address: "
+               "0x%p, len: %d\n",
+               (void *)FlashMem, GLOB_HWCTL_MEM_SIZE);
+
+       nand_dbg_print(NAND_DBG_DEBUG, "Dump timing register values:"
+                       "acc_clks: %d, re_2_we: %d, we_2_re: %d,"
+                       "addr_2_data: %d, rdwr_en_lo_cnt: %d, "
+                       "rdwr_en_hi_cnt: %d, cs_setup_cnt: %d\n",
+                       ioread32(FlashReg + ACC_CLKS),
+                       ioread32(FlashReg + RE_2_WE),
+                       ioread32(FlashReg + WE_2_RE),
+                       ioread32(FlashReg + ADDR_2_DATA),
+                       ioread32(FlashReg + RDWR_EN_LO_CNT),
+                       ioread32(FlashReg + RDWR_EN_HI_CNT),
+                       ioread32(FlashReg + CS_SETUP_CNT));
+
+       NAND_Flash_Reset();
+
+       iowrite32(0, FlashReg + GLOBAL_INT_ENABLE);
+
+#if CMD_DMA
+       info.pcmds_num = 0;
+       info.flash_bank = 0;
+       info.cdma_num = 0;
+       int_mask = (DMA_INTR__DESC_COMP_CHANNEL0 |
+               DMA_INTR__DESC_COMP_CHANNEL1 |
+               DMA_INTR__DESC_COMP_CHANNEL2 |
+               DMA_INTR__DESC_COMP_CHANNEL3 |
+               DMA_INTR__MEMCOPY_DESC_COMP);
+       iowrite32(int_mask, FlashReg + DMA_INTR_EN);
+       iowrite32(0xFFFF, FlashReg + DMA_INTR);
+
+       int_mask = (INTR_STATUS0__ECC_ERR |
+               INTR_STATUS0__PROGRAM_FAIL |
+               INTR_STATUS0__ERASE_FAIL);
+#else
+       int_mask = INTR_STATUS0__DMA_CMD_COMP |
+               INTR_STATUS0__ECC_TRANSACTION_DONE |
+               INTR_STATUS0__ECC_ERR |
+               INTR_STATUS0__PROGRAM_FAIL |
+               INTR_STATUS0__ERASE_FAIL;
+#endif
+       iowrite32(int_mask, FlashReg + INTR_EN0);
+       iowrite32(int_mask, FlashReg + INTR_EN1);
+       iowrite32(int_mask, FlashReg + INTR_EN2);
+       iowrite32(int_mask, FlashReg + INTR_EN3);
+
+       /* Clear all status bits */
+       iowrite32(0xFFFF, FlashReg + INTR_STATUS0);
+       iowrite32(0xFFFF, FlashReg + INTR_STATUS1);
+       iowrite32(0xFFFF, FlashReg + INTR_STATUS2);
+       iowrite32(0xFFFF, FlashReg + INTR_STATUS3);
+
+       iowrite32(0x0F, FlashReg + RB_PIN_ENABLED);
+       iowrite32(CHIP_EN_DONT_CARE__FLAG, FlashReg + CHIP_ENABLE_DONT_CARE);
+
+       /* Should set value for these registers when init */
+       iowrite32(0, FlashReg + TWO_ROW_ADDR_CYCLES);
+       iowrite32(1, FlashReg + ECC_ENABLE);
+       enable_ecc = 1;
+
+       retval = pci_register_driver(&nand_pci_driver);
+       if (retval)
+               return -ENOMEM;
+
+       return PASS;
+}
+
+/* Free memory */
+int nand_release(void)
+{
+       pci_unregister_driver(&nand_pci_driver);
+       iounmap(FlashMem);
+       iounmap(FlashReg);
+
+       return 0;
+}
+
+
+
diff --git a/drivers/block/spectra/lld_nand.h b/drivers/block/spectra/lld_nand.h
new file mode 100644 (file)
index 0000000..c7d62c5
--- /dev/null
@@ -0,0 +1,131 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _LLD_NAND_
+#define _LLD_NAND_
+
+#ifdef ELDORA
+#include "defs.h"
+#else
+#include "flash.h"
+#include "ffsport.h"
+#endif
+
+#define MODE_00    0x00000000
+#define MODE_01    0x04000000
+#define MODE_10    0x08000000
+#define MODE_11    0x0C000000
+
+
+#define DATA_TRANSFER_MODE              0
+#define PROTECTION_PER_BLOCK            1
+#define LOAD_WAIT_COUNT                 2
+#define PROGRAM_WAIT_COUNT              3
+#define ERASE_WAIT_COUNT                4
+#define INT_MONITOR_CYCLE_COUNT         5
+#define READ_BUSY_PIN_ENABLED           6
+#define MULTIPLANE_OPERATION_SUPPORT    7
+#define PRE_FETCH_MODE                  8
+#define CE_DONT_CARE_SUPPORT            9
+#define COPYBACK_SUPPORT                10
+#define CACHE_WRITE_SUPPORT             11
+#define CACHE_READ_SUPPORT              12
+#define NUM_PAGES_IN_BLOCK              13
+#define ECC_ENABLE_SELECT               14
+#define WRITE_ENABLE_2_READ_ENABLE      15
+#define ADDRESS_2_DATA                  16
+#define READ_ENABLE_2_WRITE_ENABLE      17
+#define TWO_ROW_ADDRESS_CYCLES          18
+#define MULTIPLANE_ADDRESS_RESTRICT     19
+#define ACC_CLOCKS                      20
+#define READ_WRITE_ENABLE_LOW_COUNT     21
+#define READ_WRITE_ENABLE_HIGH_COUNT    22
+
+#define ECC_SECTOR_SIZE     512
+#define LLD_MAX_FLASH_BANKS     4
+
+struct mrst_nand_info {
+       struct pci_dev *dev;
+       u32 state;
+       u32 flash_bank;
+       u8 *read_data;
+       u8 *write_data;
+       u32 block;
+       u16 page;
+       u32 use_dma;
+       void __iomem *ioaddr;  /* Mapped io reg base address */
+       int ret;
+       u32 pcmds_num;
+       struct pending_cmd *pcmds;
+       int cdma_num;           /* CDMA descriptor number in this chan */
+       u8 *cdma_desc_buf;      /* CDMA descriptor table */
+       u8 *memcp_desc_buf;     /* Memory copy descriptor table */
+       dma_addr_t cdma_desc;   /* Mapped CDMA descriptor table */
+       dma_addr_t memcp_desc;  /* Mapped memory copy descriptor table */
+       struct completion complete;
+};
+
+int NAND_Flash_Init(void);
+int nand_release(void);
+u16  NAND_Flash_Reset(void);
+u16  NAND_Read_Device_ID(void);
+u16  NAND_Erase_Block(u32 flash_add);
+u16  NAND_Write_Page_Main(u8 *write_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_Read_Page_Main(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_UnlockArrayAll(void);
+u16  NAND_Write_Page_Main_Spare(u8 *write_data, u32 block,
+                               u16 page, u16 page_count);
+u16  NAND_Write_Page_Spare(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_Read_Page_Main_Spare(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_Read_Page_Spare(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+void NAND_LLD_Enable_Disable_Interrupts(u16 INT_ENABLE);
+u16  NAND_Get_Bad_Block(u32 block);
+u16  NAND_Pipeline_Read_Ahead(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_Pipeline_Write_Ahead(u8 *write_data, u32 block,
+                               u16 page, u16 page_count);
+u16  NAND_Multiplane_Read(u8 *read_data, u32 block, u16 page,
+                               u16 page_count);
+u16  NAND_Multiplane_Write(u8 *write_data, u32 block, u16 page,
+                               u16 page_count);
+void NAND_ECC_Ctrl(int enable);
+u16 NAND_Read_Page_Main_Polling(u8 *read_data,
+               u32 block, u16 page, u16 page_count);
+u16 NAND_Pipeline_Read_Ahead_Polling(u8 *read_data,
+                       u32 block, u16 page, u16 page_count);
+void Conv_Spare_Data_Log2Phy_Format(u8 *data);
+void Conv_Spare_Data_Phy2Log_Format(u8 *data);
+void Conv_Main_Spare_Data_Log2Phy_Format(u8 *data, u16 page_count);
+void Conv_Main_Spare_Data_Phy2Log_Format(u8 *data, u16 page_count);
+
+extern void __iomem *FlashReg;
+extern void __iomem *FlashMem;
+
+extern int totalUsedBanks;
+extern u32 GLOB_valid_banks[LLD_MAX_FLASH_BANKS];
+
+#endif /*_LLD_NAND_*/
+
+
+
diff --git a/drivers/block/spectra/nand_regs.h b/drivers/block/spectra/nand_regs.h
new file mode 100644 (file)
index 0000000..e192e4a
--- /dev/null
@@ -0,0 +1,619 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#define DEVICE_RESET                           0x0
+#define     DEVICE_RESET__BANK0                                0x0001
+#define     DEVICE_RESET__BANK1                                0x0002
+#define     DEVICE_RESET__BANK2                                0x0004
+#define     DEVICE_RESET__BANK3                                0x0008
+
+#define TRANSFER_SPARE_REG                     0x10
+#define     TRANSFER_SPARE_REG__FLAG                   0x0001
+
+#define LOAD_WAIT_CNT                          0x20
+#define     LOAD_WAIT_CNT__VALUE                               0xffff
+
+#define PROGRAM_WAIT_CNT                       0x30
+#define     PROGRAM_WAIT_CNT__VALUE                    0xffff
+
+#define ERASE_WAIT_CNT                         0x40
+#define     ERASE_WAIT_CNT__VALUE                      0xffff
+
+#define INT_MON_CYCCNT                         0x50
+#define     INT_MON_CYCCNT__VALUE                      0xffff
+
+#define RB_PIN_ENABLED                         0x60
+#define     RB_PIN_ENABLED__BANK0                      0x0001
+#define     RB_PIN_ENABLED__BANK1                      0x0002
+#define     RB_PIN_ENABLED__BANK2                      0x0004
+#define     RB_PIN_ENABLED__BANK3                      0x0008
+
+#define MULTIPLANE_OPERATION                   0x70
+#define     MULTIPLANE_OPERATION__FLAG                 0x0001
+
+#define MULTIPLANE_READ_ENABLE                 0x80
+#define     MULTIPLANE_READ_ENABLE__FLAG               0x0001
+
+#define COPYBACK_DISABLE                       0x90
+#define     COPYBACK_DISABLE__FLAG                     0x0001
+
+#define CACHE_WRITE_ENABLE                     0xa0
+#define     CACHE_WRITE_ENABLE__FLAG                   0x0001
+
+#define CACHE_READ_ENABLE                      0xb0
+#define     CACHE_READ_ENABLE__FLAG                    0x0001
+
+#define PREFETCH_MODE                          0xc0
+#define     PREFETCH_MODE__PREFETCH_EN                 0x0001
+#define     PREFETCH_MODE__PREFETCH_BURST_LENGTH       0xfff0
+
+#define CHIP_ENABLE_DONT_CARE                  0xd0
+#define     CHIP_EN_DONT_CARE__FLAG                    0x01
+
+#define ECC_ENABLE                             0xe0
+#define     ECC_ENABLE__FLAG                           0x0001
+
+#define GLOBAL_INT_ENABLE                      0xf0
+#define     GLOBAL_INT_EN_FLAG                         0x01
+
+#define WE_2_RE                                        0x100
+#define     WE_2_RE__VALUE                             0x003f
+
+#define ADDR_2_DATA                            0x110
+#define     ADDR_2_DATA__VALUE                         0x003f
+
+#define RE_2_WE                                        0x120
+#define     RE_2_WE__VALUE                             0x003f
+
+#define ACC_CLKS                               0x130
+#define     ACC_CLKS__VALUE                            0x000f
+
+#define NUMBER_OF_PLANES                       0x140
+#define     NUMBER_OF_PLANES__VALUE                    0x0007
+
+#define PAGES_PER_BLOCK                                0x150
+#define     PAGES_PER_BLOCK__VALUE                     0xffff
+
+#define DEVICE_WIDTH                           0x160
+#define     DEVICE_WIDTH__VALUE                                0x0003
+
+#define DEVICE_MAIN_AREA_SIZE                  0x170
+#define     DEVICE_MAIN_AREA_SIZE__VALUE               0xffff
+
+#define DEVICE_SPARE_AREA_SIZE                 0x180
+#define     DEVICE_SPARE_AREA_SIZE__VALUE              0xffff
+
+#define TWO_ROW_ADDR_CYCLES                    0x190
+#define     TWO_ROW_ADDR_CYCLES__FLAG                  0x0001
+
+#define MULTIPLANE_ADDR_RESTRICT               0x1a0
+#define     MULTIPLANE_ADDR_RESTRICT__FLAG             0x0001
+
+#define ECC_CORRECTION                         0x1b0
+#define     ECC_CORRECTION__VALUE                      0x001f
+
+#define READ_MODE                              0x1c0
+#define     READ_MODE__VALUE                           0x000f
+
+#define WRITE_MODE                             0x1d0
+#define     WRITE_MODE__VALUE                          0x000f
+
+#define COPYBACK_MODE                          0x1e0
+#define     COPYBACK_MODE__VALUE                       0x000f
+
+#define RDWR_EN_LO_CNT                         0x1f0
+#define     RDWR_EN_LO_CNT__VALUE                      0x001f
+
+#define RDWR_EN_HI_CNT                         0x200
+#define     RDWR_EN_HI_CNT__VALUE                      0x001f
+
+#define MAX_RD_DELAY                           0x210
+#define     MAX_RD_DELAY__VALUE                                0x000f
+
+#define CS_SETUP_CNT                           0x220
+#define     CS_SETUP_CNT__VALUE                                0x001f
+
+#define SPARE_AREA_SKIP_BYTES                  0x230
+#define     SPARE_AREA_SKIP_BYTES__VALUE               0x003f
+
+#define SPARE_AREA_MARKER                      0x240
+#define     SPARE_AREA_MARKER__VALUE                   0xffff
+
+#define DEVICES_CONNECTED                      0x250
+#define     DEVICES_CONNECTED__VALUE                   0x0007
+
+#define DIE_MASK                                       0x260
+#define     DIE_MASK__VALUE                            0x00ff
+
+#define FIRST_BLOCK_OF_NEXT_PLANE              0x270
+#define     FIRST_BLOCK_OF_NEXT_PLANE__VALUE           0xffff
+
+#define WRITE_PROTECT                          0x280
+#define     WRITE_PROTECT__FLAG                                0x0001
+
+#define RE_2_RE                                        0x290
+#define     RE_2_RE__VALUE                             0x003f
+
+#define MANUFACTURER_ID                        0x300
+#define     MANUFACTURER_ID__VALUE                     0x00ff
+
+#define DEVICE_ID                              0x310
+#define     DEVICE_ID__VALUE                           0x00ff
+
+#define DEVICE_PARAM_0                         0x320
+#define     DEVICE_PARAM_0__VALUE                      0x00ff
+
+#define DEVICE_PARAM_1                         0x330
+#define     DEVICE_PARAM_1__VALUE                      0x00ff
+
+#define DEVICE_PARAM_2                         0x340
+#define     DEVICE_PARAM_2__VALUE                      0x00ff
+
+#define LOGICAL_PAGE_DATA_SIZE                 0x350
+#define     LOGICAL_PAGE_DATA_SIZE__VALUE              0xffff
+
+#define LOGICAL_PAGE_SPARE_SIZE                        0x360
+#define     LOGICAL_PAGE_SPARE_SIZE__VALUE             0xffff
+
+#define REVISION                                       0x370
+#define     REVISION__VALUE                            0xffff
+
+#define ONFI_DEVICE_FEATURES                   0x380
+#define     ONFI_DEVICE_FEATURES__VALUE                        0x003f
+
+#define ONFI_OPTIONAL_COMMANDS         0x390
+#define     ONFI_OPTIONAL_COMMANDS__VALUE              0x003f
+
+#define ONFI_TIMING_MODE                       0x3a0
+#define     ONFI_TIMING_MODE__VALUE                    0x003f
+
+#define ONFI_PGM_CACHE_TIMING_MODE             0x3b0
+#define     ONFI_PGM_CACHE_TIMING_MODE__VALUE          0x003f
+
+#define ONFI_DEVICE_NO_OF_LUNS                 0x3c0
+#define     ONFI_DEVICE_NO_OF_LUNS__NO_OF_LUNS         0x00ff
+#define     ONFI_DEVICE_NO_OF_LUNS__ONFI_DEVICE                0x0100
+
+#define ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_L     0x3d0
+#define     ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_L__VALUE  0xffff
+
+#define ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_U     0x3e0
+#define     ONFI_DEVICE_NO_OF_BLOCKS_PER_LUN_U__VALUE  0xffff
+
+#define FEATURES                                       0x3f0
+#define     FEATURES__N_BANKS                          0x0003
+#define     FEATURES__ECC_MAX_ERR                      0x003c
+#define     FEATURES__DMA                                      0x0040
+#define     FEATURES__CMD_DMA                          0x0080
+#define     FEATURES__PARTITION                                0x0100
+#define     FEATURES__XDMA_SIDEBAND                    0x0200
+#define     FEATURES__GPREG                            0x0400
+#define     FEATURES__INDEX_ADDR                               0x0800
+
+#define TRANSFER_MODE                          0x400
+#define     TRANSFER_MODE__VALUE                       0x0003
+
+#define INTR_STATUS0                           0x410
+#define     INTR_STATUS0__ECC_TRANSACTION_DONE         0x0001
+#define     INTR_STATUS0__ECC_ERR                      0x0002
+#define     INTR_STATUS0__DMA_CMD_COMP                 0x0004
+#define     INTR_STATUS0__TIME_OUT                     0x0008
+#define     INTR_STATUS0__PROGRAM_FAIL                 0x0010
+#define     INTR_STATUS0__ERASE_FAIL                   0x0020
+#define     INTR_STATUS0__LOAD_COMP                    0x0040
+#define     INTR_STATUS0__PROGRAM_COMP                 0x0080
+#define     INTR_STATUS0__ERASE_COMP                   0x0100
+#define     INTR_STATUS0__PIPE_CPYBCK_CMD_COMP         0x0200
+#define     INTR_STATUS0__LOCKED_BLK                   0x0400
+#define     INTR_STATUS0__UNSUP_CMD                    0x0800
+#define     INTR_STATUS0__INT_ACT                      0x1000
+#define     INTR_STATUS0__RST_COMP                     0x2000
+#define     INTR_STATUS0__PIPE_CMD_ERR                 0x4000
+#define     INTR_STATUS0__PAGE_XFER_INC                        0x8000
+
+#define INTR_EN0                                       0x420
+#define     INTR_EN0__ECC_TRANSACTION_DONE             0x0001
+#define     INTR_EN0__ECC_ERR                          0x0002
+#define     INTR_EN0__DMA_CMD_COMP                     0x0004
+#define     INTR_EN0__TIME_OUT                         0x0008
+#define     INTR_EN0__PROGRAM_FAIL                     0x0010
+#define     INTR_EN0__ERASE_FAIL                               0x0020
+#define     INTR_EN0__LOAD_COMP                                0x0040
+#define     INTR_EN0__PROGRAM_COMP                     0x0080
+#define     INTR_EN0__ERASE_COMP                               0x0100
+#define     INTR_EN0__PIPE_CPYBCK_CMD_COMP             0x0200
+#define     INTR_EN0__LOCKED_BLK                               0x0400
+#define     INTR_EN0__UNSUP_CMD                                0x0800
+#define     INTR_EN0__INT_ACT                          0x1000
+#define     INTR_EN0__RST_COMP                         0x2000
+#define     INTR_EN0__PIPE_CMD_ERR                     0x4000
+#define     INTR_EN0__PAGE_XFER_INC                    0x8000
+
+#define PAGE_CNT0                              0x430
+#define     PAGE_CNT0__VALUE                           0x00ff
+
+#define ERR_PAGE_ADDR0                         0x440
+#define     ERR_PAGE_ADDR0__VALUE                      0xffff
+
+#define ERR_BLOCK_ADDR0                        0x450
+#define     ERR_BLOCK_ADDR0__VALUE                     0xffff
+
+#define INTR_STATUS1                           0x460
+#define     INTR_STATUS1__ECC_TRANSACTION_DONE         0x0001
+#define     INTR_STATUS1__ECC_ERR                      0x0002
+#define     INTR_STATUS1__DMA_CMD_COMP                 0x0004
+#define     INTR_STATUS1__TIME_OUT                     0x0008
+#define     INTR_STATUS1__PROGRAM_FAIL                 0x0010
+#define     INTR_STATUS1__ERASE_FAIL                   0x0020
+#define     INTR_STATUS1__LOAD_COMP                    0x0040
+#define     INTR_STATUS1__PROGRAM_COMP                 0x0080
+#define     INTR_STATUS1__ERASE_COMP                   0x0100
+#define     INTR_STATUS1__PIPE_CPYBCK_CMD_COMP         0x0200
+#define     INTR_STATUS1__LOCKED_BLK                   0x0400
+#define     INTR_STATUS1__UNSUP_CMD                    0x0800
+#define     INTR_STATUS1__INT_ACT                      0x1000
+#define     INTR_STATUS1__RST_COMP                     0x2000
+#define     INTR_STATUS1__PIPE_CMD_ERR                 0x4000
+#define     INTR_STATUS1__PAGE_XFER_INC                        0x8000
+
+#define INTR_EN1                                       0x470
+#define     INTR_EN1__ECC_TRANSACTION_DONE             0x0001
+#define     INTR_EN1__ECC_ERR                          0x0002
+#define     INTR_EN1__DMA_CMD_COMP                     0x0004
+#define     INTR_EN1__TIME_OUT                         0x0008
+#define     INTR_EN1__PROGRAM_FAIL                     0x0010
+#define     INTR_EN1__ERASE_FAIL                               0x0020
+#define     INTR_EN1__LOAD_COMP                                0x0040
+#define     INTR_EN1__PROGRAM_COMP                     0x0080
+#define     INTR_EN1__ERASE_COMP                               0x0100
+#define     INTR_EN1__PIPE_CPYBCK_CMD_COMP             0x0200
+#define     INTR_EN1__LOCKED_BLK                               0x0400
+#define     INTR_EN1__UNSUP_CMD                                0x0800
+#define     INTR_EN1__INT_ACT                          0x1000
+#define     INTR_EN1__RST_COMP                         0x2000
+#define     INTR_EN1__PIPE_CMD_ERR                     0x4000
+#define     INTR_EN1__PAGE_XFER_INC                    0x8000
+
+#define PAGE_CNT1                              0x480
+#define     PAGE_CNT1__VALUE                           0x00ff
+
+#define ERR_PAGE_ADDR1                         0x490
+#define     ERR_PAGE_ADDR1__VALUE                      0xffff
+
+#define ERR_BLOCK_ADDR1                        0x4a0
+#define     ERR_BLOCK_ADDR1__VALUE                     0xffff
+
+#define INTR_STATUS2                           0x4b0
+#define     INTR_STATUS2__ECC_TRANSACTION_DONE         0x0001
+#define     INTR_STATUS2__ECC_ERR                      0x0002
+#define     INTR_STATUS2__DMA_CMD_COMP                 0x0004
+#define     INTR_STATUS2__TIME_OUT                     0x0008
+#define     INTR_STATUS2__PROGRAM_FAIL                 0x0010
+#define     INTR_STATUS2__ERASE_FAIL                   0x0020
+#define     INTR_STATUS2__LOAD_COMP                    0x0040
+#define     INTR_STATUS2__PROGRAM_COMP                 0x0080
+#define     INTR_STATUS2__ERASE_COMP                   0x0100
+#define     INTR_STATUS2__PIPE_CPYBCK_CMD_COMP         0x0200
+#define     INTR_STATUS2__LOCKED_BLK                   0x0400
+#define     INTR_STATUS2__UNSUP_CMD                    0x0800
+#define     INTR_STATUS2__INT_ACT                      0x1000
+#define     INTR_STATUS2__RST_COMP                     0x2000
+#define     INTR_STATUS2__PIPE_CMD_ERR                 0x4000
+#define     INTR_STATUS2__PAGE_XFER_INC                        0x8000
+
+#define INTR_EN2                                       0x4c0
+#define     INTR_EN2__ECC_TRANSACTION_DONE             0x0001
+#define     INTR_EN2__ECC_ERR                          0x0002
+#define     INTR_EN2__DMA_CMD_COMP                     0x0004
+#define     INTR_EN2__TIME_OUT                         0x0008
+#define     INTR_EN2__PROGRAM_FAIL                     0x0010
+#define     INTR_EN2__ERASE_FAIL                               0x0020
+#define     INTR_EN2__LOAD_COMP                                0x0040
+#define     INTR_EN2__PROGRAM_COMP                     0x0080
+#define     INTR_EN2__ERASE_COMP                               0x0100
+#define     INTR_EN2__PIPE_CPYBCK_CMD_COMP             0x0200
+#define     INTR_EN2__LOCKED_BLK                               0x0400
+#define     INTR_EN2__UNSUP_CMD                                0x0800
+#define     INTR_EN2__INT_ACT                          0x1000
+#define     INTR_EN2__RST_COMP                         0x2000
+#define     INTR_EN2__PIPE_CMD_ERR                     0x4000
+#define     INTR_EN2__PAGE_XFER_INC                    0x8000
+
+#define PAGE_CNT2                              0x4d0
+#define     PAGE_CNT2__VALUE                           0x00ff
+
+#define ERR_PAGE_ADDR2                         0x4e0
+#define     ERR_PAGE_ADDR2__VALUE                      0xffff
+
+#define ERR_BLOCK_ADDR2                        0x4f0
+#define     ERR_BLOCK_ADDR2__VALUE                     0xffff
+
+#define INTR_STATUS3                           0x500
+#define     INTR_STATUS3__ECC_TRANSACTION_DONE         0x0001
+#define     INTR_STATUS3__ECC_ERR                      0x0002
+#define     INTR_STATUS3__DMA_CMD_COMP                 0x0004
+#define     INTR_STATUS3__TIME_OUT                     0x0008
+#define     INTR_STATUS3__PROGRAM_FAIL                 0x0010
+#define     INTR_STATUS3__ERASE_FAIL                   0x0020
+#define     INTR_STATUS3__LOAD_COMP                    0x0040
+#define     INTR_STATUS3__PROGRAM_COMP                 0x0080
+#define     INTR_STATUS3__ERASE_COMP                   0x0100
+#define     INTR_STATUS3__PIPE_CPYBCK_CMD_COMP         0x0200
+#define     INTR_STATUS3__LOCKED_BLK                   0x0400
+#define     INTR_STATUS3__UNSUP_CMD                    0x0800
+#define     INTR_STATUS3__INT_ACT                      0x1000
+#define     INTR_STATUS3__RST_COMP                     0x2000
+#define     INTR_STATUS3__PIPE_CMD_ERR                 0x4000
+#define     INTR_STATUS3__PAGE_XFER_INC                        0x8000
+
+#define INTR_EN3                                       0x510
+#define     INTR_EN3__ECC_TRANSACTION_DONE             0x0001
+#define     INTR_EN3__ECC_ERR                          0x0002
+#define     INTR_EN3__DMA_CMD_COMP                     0x0004
+#define     INTR_EN3__TIME_OUT                         0x0008
+#define     INTR_EN3__PROGRAM_FAIL                     0x0010
+#define     INTR_EN3__ERASE_FAIL                               0x0020
+#define     INTR_EN3__LOAD_COMP                                0x0040
+#define     INTR_EN3__PROGRAM_COMP                     0x0080
+#define     INTR_EN3__ERASE_COMP                               0x0100
+#define     INTR_EN3__PIPE_CPYBCK_CMD_COMP             0x0200
+#define     INTR_EN3__LOCKED_BLK                               0x0400
+#define     INTR_EN3__UNSUP_CMD                                0x0800
+#define     INTR_EN3__INT_ACT                          0x1000
+#define     INTR_EN3__RST_COMP                         0x2000
+#define     INTR_EN3__PIPE_CMD_ERR                     0x4000
+#define     INTR_EN3__PAGE_XFER_INC                    0x8000
+
+#define PAGE_CNT3                              0x520
+#define     PAGE_CNT3__VALUE                           0x00ff
+
+#define ERR_PAGE_ADDR3                         0x530
+#define     ERR_PAGE_ADDR3__VALUE                      0xffff
+
+#define ERR_BLOCK_ADDR3                        0x540
+#define     ERR_BLOCK_ADDR3__VALUE                     0xffff
+
+#define DATA_INTR                              0x550
+#define     DATA_INTR__WRITE_SPACE_AV                  0x0001
+#define     DATA_INTR__READ_DATA_AV                    0x0002
+
+#define DATA_INTR_EN                           0x560
+#define     DATA_INTR_EN__WRITE_SPACE_AV               0x0001
+#define     DATA_INTR_EN__READ_DATA_AV                 0x0002
+
+#define GPREG_0                                        0x570
+#define     GPREG_0__VALUE                             0xffff
+
+#define GPREG_1                                        0x580
+#define     GPREG_1__VALUE                             0xffff
+
+#define GPREG_2                                        0x590
+#define     GPREG_2__VALUE                             0xffff
+
+#define GPREG_3                                        0x5a0
+#define     GPREG_3__VALUE                             0xffff
+
+#define ECC_THRESHOLD                          0x600
+#define     ECC_THRESHOLD__VALUE                               0x03ff
+
+#define ECC_ERROR_BLOCK_ADDRESS                0x610
+#define     ECC_ERROR_BLOCK_ADDRESS__VALUE             0xffff
+
+#define ECC_ERROR_PAGE_ADDRESS                 0x620
+#define     ECC_ERROR_PAGE_ADDRESS__VALUE              0x0fff
+#define     ECC_ERROR_PAGE_ADDRESS__BANK               0xf000
+
+#define ECC_ERROR_ADDRESS                      0x630
+#define     ECC_ERROR_ADDRESS__OFFSET                  0x0fff
+#define     ECC_ERROR_ADDRESS__SECTOR_NR               0xf000
+
+#define ERR_CORRECTION_INFO                    0x640
+#define     ERR_CORRECTION_INFO__BYTEMASK              0x00ff
+#define     ERR_CORRECTION_INFO__DEVICE_NR             0x0f00
+#define     ERR_CORRECTION_INFO__ERROR_TYPE            0x4000
+#define     ERR_CORRECTION_INFO__LAST_ERR_INFO         0x8000
+
+#define DMA_ENABLE                             0x700
+#define     DMA_ENABLE__FLAG                           0x0001
+
+#define IGNORE_ECC_DONE                                0x710
+#define     IGNORE_ECC_DONE__FLAG                      0x0001
+
+#define DMA_INTR                               0x720
+#define     DMA_INTR__TARGET_ERROR                     0x0001
+#define     DMA_INTR__DESC_COMP_CHANNEL0               0x0002
+#define     DMA_INTR__DESC_COMP_CHANNEL1               0x0004
+#define     DMA_INTR__DESC_COMP_CHANNEL2               0x0008
+#define     DMA_INTR__DESC_COMP_CHANNEL3               0x0010
+#define     DMA_INTR__MEMCOPY_DESC_COMP                0x0020
+
+#define DMA_INTR_EN                            0x730
+#define     DMA_INTR_EN__TARGET_ERROR                  0x0001
+#define     DMA_INTR_EN__DESC_COMP_CHANNEL0            0x0002
+#define     DMA_INTR_EN__DESC_COMP_CHANNEL1            0x0004
+#define     DMA_INTR_EN__DESC_COMP_CHANNEL2            0x0008
+#define     DMA_INTR_EN__DESC_COMP_CHANNEL3            0x0010
+#define     DMA_INTR_EN__MEMCOPY_DESC_COMP             0x0020
+
+#define TARGET_ERR_ADDR_LO                     0x740
+#define     TARGET_ERR_ADDR_LO__VALUE                  0xffff
+
+#define TARGET_ERR_ADDR_HI                     0x750
+#define     TARGET_ERR_ADDR_HI__VALUE                  0xffff
+
+#define CHNL_ACTIVE                            0x760
+#define     CHNL_ACTIVE__CHANNEL0                      0x0001
+#define     CHNL_ACTIVE__CHANNEL1                      0x0002
+#define     CHNL_ACTIVE__CHANNEL2                      0x0004
+#define     CHNL_ACTIVE__CHANNEL3                      0x0008
+
+#define ACTIVE_SRC_ID                          0x800
+#define     ACTIVE_SRC_ID__VALUE                               0x00ff
+
+#define PTN_INTR                                       0x810
+#define     PTN_INTR__CONFIG_ERROR                     0x0001
+#define     PTN_INTR__ACCESS_ERROR_BANK0               0x0002
+#define     PTN_INTR__ACCESS_ERROR_BANK1               0x0004
+#define     PTN_INTR__ACCESS_ERROR_BANK2               0x0008
+#define     PTN_INTR__ACCESS_ERROR_BANK3               0x0010
+#define     PTN_INTR__REG_ACCESS_ERROR                 0x0020
+
+#define PTN_INTR_EN                            0x820
+#define     PTN_INTR_EN__CONFIG_ERROR                  0x0001
+#define     PTN_INTR_EN__ACCESS_ERROR_BANK0            0x0002
+#define     PTN_INTR_EN__ACCESS_ERROR_BANK1            0x0004
+#define     PTN_INTR_EN__ACCESS_ERROR_BANK2            0x0008
+#define     PTN_INTR_EN__ACCESS_ERROR_BANK3            0x0010
+#define     PTN_INTR_EN__REG_ACCESS_ERROR              0x0020
+
+#define PERM_SRC_ID_0                          0x830
+#define     PERM_SRC_ID_0__SRCID                               0x00ff
+#define     PERM_SRC_ID_0__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_0__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_0__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_0__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_0                         0x840
+#define     MIN_BLK_ADDR_0__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_0                         0x850
+#define     MAX_BLK_ADDR_0__VALUE                      0xffff
+
+#define MIN_MAX_BANK_0                         0x860
+#define     MIN_MAX_BANK_0__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_0__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_1                          0x870
+#define     PERM_SRC_ID_1__SRCID                               0x00ff
+#define     PERM_SRC_ID_1__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_1__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_1__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_1__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_1                         0x880
+#define     MIN_BLK_ADDR_1__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_1                         0x890
+#define     MAX_BLK_ADDR_1__VALUE                      0xffff
+
+#define MIN_MAX_BANK_1                         0x8a0
+#define     MIN_MAX_BANK_1__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_1__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_2                          0x8b0
+#define     PERM_SRC_ID_2__SRCID                               0x00ff
+#define     PERM_SRC_ID_2__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_2__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_2__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_2__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_2                         0x8c0
+#define     MIN_BLK_ADDR_2__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_2                         0x8d0
+#define     MAX_BLK_ADDR_2__VALUE                      0xffff
+
+#define MIN_MAX_BANK_2                         0x8e0
+#define     MIN_MAX_BANK_2__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_2__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_3                          0x8f0
+#define     PERM_SRC_ID_3__SRCID                               0x00ff
+#define     PERM_SRC_ID_3__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_3__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_3__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_3__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_3                         0x900
+#define     MIN_BLK_ADDR_3__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_3                         0x910
+#define     MAX_BLK_ADDR_3__VALUE                      0xffff
+
+#define MIN_MAX_BANK_3                         0x920
+#define     MIN_MAX_BANK_3__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_3__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_4                          0x930
+#define     PERM_SRC_ID_4__SRCID                               0x00ff
+#define     PERM_SRC_ID_4__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_4__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_4__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_4__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_4                         0x940
+#define     MIN_BLK_ADDR_4__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_4                         0x950
+#define     MAX_BLK_ADDR_4__VALUE                      0xffff
+
+#define MIN_MAX_BANK_4                         0x960
+#define     MIN_MAX_BANK_4__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_4__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_5                          0x970
+#define     PERM_SRC_ID_5__SRCID                               0x00ff
+#define     PERM_SRC_ID_5__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_5__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_5__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_5__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_5                         0x980
+#define     MIN_BLK_ADDR_5__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_5                         0x990
+#define     MAX_BLK_ADDR_5__VALUE                      0xffff
+
+#define MIN_MAX_BANK_5                         0x9a0
+#define     MIN_MAX_BANK_5__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_5__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_6                          0x9b0
+#define     PERM_SRC_ID_6__SRCID                               0x00ff
+#define     PERM_SRC_ID_6__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_6__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_6__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_6__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_6                         0x9c0
+#define     MIN_BLK_ADDR_6__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_6                         0x9d0
+#define     MAX_BLK_ADDR_6__VALUE                      0xffff
+
+#define MIN_MAX_BANK_6                         0x9e0
+#define     MIN_MAX_BANK_6__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_6__MAX_VALUE                  0x000c
+
+#define PERM_SRC_ID_7                          0x9f0
+#define     PERM_SRC_ID_7__SRCID                               0x00ff
+#define     PERM_SRC_ID_7__DIRECT_ACCESS_ACTIVE                0x0800
+#define     PERM_SRC_ID_7__WRITE_ACTIVE                        0x2000
+#define     PERM_SRC_ID_7__READ_ACTIVE                 0x4000
+#define     PERM_SRC_ID_7__PARTITION_VALID             0x8000
+
+#define MIN_BLK_ADDR_7                         0xa00
+#define     MIN_BLK_ADDR_7__VALUE                      0xffff
+
+#define MAX_BLK_ADDR_7                         0xa10
+#define     MAX_BLK_ADDR_7__VALUE                      0xffff
+
+#define MIN_MAX_BANK_7                         0xa20
+#define     MIN_MAX_BANK_7__MIN_VALUE                  0x0003
+#define     MIN_MAX_BANK_7__MAX_VALUE                  0x000c
diff --git a/drivers/block/spectra/spectraswconfig.h b/drivers/block/spectra/spectraswconfig.h
new file mode 100644 (file)
index 0000000..b630f06
--- /dev/null
@@ -0,0 +1,81 @@
+/*
+ * NAND Flash Controller Device Driver
+ * Copyright (c) 2009, Intel Corporation and its suppliers.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms and conditions of the GNU General Public License,
+ * version 2, as published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ *
+ * You should have received a copy of the GNU General Public License along with
+ * this program; if not, write to the Free Software Foundation, Inc.,
+ * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ */
+
+#ifndef _SPECTRASWCONFIG_
+#define _SPECTRASWCONFIG_
+
+/* NAND driver version */
+#define GLOB_VERSION          "driver version 20100311"
+
+
+/***** Common Parameters *****/
+#define RETRY_TIMES                   3
+
+#define READ_BADBLOCK_INFO            1
+#define READBACK_VERIFY               0
+#define AUTO_FORMAT_FLASH             0
+
+/***** Cache Parameters *****/
+#define CACHE_ITEM_NUM            128
+#define BLK_NUM_FOR_L2_CACHE        16
+
+/***** Block Table Parameters *****/
+#define BLOCK_TABLE_INDEX             0
+
+/***** Wear Leveling Parameters *****/
+#define WEAR_LEVELING_GATE         0x10
+#define WEAR_LEVELING_BLOCK_NUM      10
+
+#define DEBUG_BNDRY             0
+
+/***** Product Feature Support *****/
+#define FLASH_EMU               defined(CONFIG_MRST_NAND_EMU)
+#define FLASH_NAND              defined(CONFIG_MRST_NAND_HW)
+#define CMD_DMA                   0
+
+#define SPECTRA_PARTITION_ID    0
+
+/* Enable this macro if the number of flash blocks is larger than 16K. */
+#define SUPPORT_LARGE_BLOCKNUM  1
+
+/**** Block Table and Reserved Block Parameters *****/
+#define SPECTRA_START_BLOCK     3
+//#define NUM_FREE_BLOCKS_GATE    30
+#define NUM_FREE_BLOCKS_GATE    60
+
+/**** Hardware Parameters ****/
+#define GLOB_HWCTL_REG_BASE     0xFFA40000
+#define GLOB_HWCTL_REG_SIZE     4096
+
+#define GLOB_HWCTL_MEM_BASE     0xFFA48000
+#define GLOB_HWCTL_MEM_SIZE     4096
+
+/* KBV - Updated to LNW scratch register address */
+#define SCRATCH_REG_ADDR    0xFF108018
+#define SCRATCH_REG_SIZE    64
+
+#define GLOB_HWCTL_DEFAULT_BLKS    2048
+
+#define SUPPORT_15BITECC        1
+#define SUPPORT_8BITECC         1
+
+#define ONFI_BLOOM_TIME         0
+#define MODE5_WORKAROUND        1
+
+#endif /*_SPECTRASWCONFIG_*/