1*4882a593Smuzhiyun /* This version ported to the Linux-MTD system by dwmw2@infradead.org
2*4882a593Smuzhiyun *
3*4882a593Smuzhiyun * Fixes: Arnaldo Carvalho de Melo <acme@conectiva.com.br>
4*4882a593Smuzhiyun * - fixes some leaks on failure in build_maps and ftl_notify_add, cleanups
5*4882a593Smuzhiyun *
6*4882a593Smuzhiyun * Based on:
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun /*======================================================================
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun A Flash Translation Layer memory card driver
11*4882a593Smuzhiyun
12*4882a593Smuzhiyun This driver implements a disk-like block device driver with an
13*4882a593Smuzhiyun apparent block size of 512 bytes for flash memory cards.
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun ftl_cs.c 1.62 2000/02/01 00:59:04
16*4882a593Smuzhiyun
17*4882a593Smuzhiyun The contents of this file are subject to the Mozilla Public
18*4882a593Smuzhiyun License Version 1.1 (the "License"); you may not use this file
19*4882a593Smuzhiyun except in compliance with the License. You may obtain a copy of
20*4882a593Smuzhiyun the License at http://www.mozilla.org/MPL/
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun Software distributed under the License is distributed on an "AS
23*4882a593Smuzhiyun IS" basis, WITHOUT WARRANTY OF ANY KIND, either express or
24*4882a593Smuzhiyun implied. See the License for the specific language governing
25*4882a593Smuzhiyun rights and limitations under the License.
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun The initial developer of the original code is David A. Hinds
28*4882a593Smuzhiyun <dahinds@users.sourceforge.net>. Portions created by David A. Hinds
29*4882a593Smuzhiyun are Copyright © 1999 David A. Hinds. All Rights Reserved.
30*4882a593Smuzhiyun
31*4882a593Smuzhiyun Alternatively, the contents of this file may be used under the
32*4882a593Smuzhiyun terms of the GNU General Public License version 2 (the "GPL"), in
33*4882a593Smuzhiyun which case the provisions of the GPL are applicable instead of the
34*4882a593Smuzhiyun above. If you wish to allow the use of your version of this file
35*4882a593Smuzhiyun only under the terms of the GPL and not to allow others to use
36*4882a593Smuzhiyun your version of this file under the MPL, indicate your decision
37*4882a593Smuzhiyun by deleting the provisions above and replace them with the notice
38*4882a593Smuzhiyun and other provisions required by the GPL. If you do not delete
39*4882a593Smuzhiyun the provisions above, a recipient may use your version of this
40*4882a593Smuzhiyun file under either the MPL or the GPL.
41*4882a593Smuzhiyun
42*4882a593Smuzhiyun LEGAL NOTE: The FTL format is patented by M-Systems. They have
43*4882a593Smuzhiyun granted a license for its use with PCMCIA devices:
44*4882a593Smuzhiyun
45*4882a593Smuzhiyun "M-Systems grants a royalty-free, non-exclusive license under
46*4882a593Smuzhiyun any presently existing M-Systems intellectual property rights
47*4882a593Smuzhiyun necessary for the design and development of FTL-compatible
48*4882a593Smuzhiyun drivers, file systems and utilities using the data formats with
49*4882a593Smuzhiyun PCMCIA PC Cards as described in the PCMCIA Flash Translation
50*4882a593Smuzhiyun Layer (FTL) Specification."
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun Use of the FTL format for non-PCMCIA applications may be an
53*4882a593Smuzhiyun infringement of these patents. For additional information,
54*4882a593Smuzhiyun contact M-Systems directly. M-Systems since acquired by Sandisk.
55*4882a593Smuzhiyun
56*4882a593Smuzhiyun ======================================================================*/
57*4882a593Smuzhiyun #include <linux/mtd/blktrans.h>
58*4882a593Smuzhiyun #include <linux/module.h>
59*4882a593Smuzhiyun #include <linux/mtd/mtd.h>
60*4882a593Smuzhiyun /*#define PSYCHO_DEBUG */
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun #include <linux/kernel.h>
63*4882a593Smuzhiyun #include <linux/ptrace.h>
64*4882a593Smuzhiyun #include <linux/slab.h>
65*4882a593Smuzhiyun #include <linux/string.h>
66*4882a593Smuzhiyun #include <linux/timer.h>
67*4882a593Smuzhiyun #include <linux/major.h>
68*4882a593Smuzhiyun #include <linux/fs.h>
69*4882a593Smuzhiyun #include <linux/init.h>
70*4882a593Smuzhiyun #include <linux/hdreg.h>
71*4882a593Smuzhiyun #include <linux/vmalloc.h>
72*4882a593Smuzhiyun #include <linux/blkpg.h>
73*4882a593Smuzhiyun #include <linux/uaccess.h>
74*4882a593Smuzhiyun
75*4882a593Smuzhiyun #include <linux/mtd/ftl.h>
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun /*====================================================================*/
78*4882a593Smuzhiyun
79*4882a593Smuzhiyun /* Parameters that can be set with 'insmod' */
80*4882a593Smuzhiyun static int shuffle_freq = 50;
81*4882a593Smuzhiyun module_param(shuffle_freq, int, 0);
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun /*====================================================================*/
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun /* Major device # for FTL device */
86*4882a593Smuzhiyun #ifndef FTL_MAJOR
87*4882a593Smuzhiyun #define FTL_MAJOR 44
88*4882a593Smuzhiyun #endif
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun /*====================================================================*/
92*4882a593Smuzhiyun
93*4882a593Smuzhiyun /* Maximum number of separate memory devices we'll allow */
94*4882a593Smuzhiyun #define MAX_DEV 4
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun /* Maximum number of regions per device */
97*4882a593Smuzhiyun #define MAX_REGION 4
98*4882a593Smuzhiyun
99*4882a593Smuzhiyun /* Maximum number of partitions in an FTL region */
100*4882a593Smuzhiyun #define PART_BITS 4
101*4882a593Smuzhiyun
102*4882a593Smuzhiyun /* Maximum number of outstanding erase requests per socket */
103*4882a593Smuzhiyun #define MAX_ERASE 8
104*4882a593Smuzhiyun
105*4882a593Smuzhiyun /* Sector size -- shouldn't need to change */
106*4882a593Smuzhiyun #define SECTOR_SIZE 512
107*4882a593Smuzhiyun
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun /* Each memory region corresponds to a minor device */
110*4882a593Smuzhiyun typedef struct partition_t {
111*4882a593Smuzhiyun struct mtd_blktrans_dev mbd;
112*4882a593Smuzhiyun uint32_t state;
113*4882a593Smuzhiyun uint32_t *VirtualBlockMap;
114*4882a593Smuzhiyun uint32_t FreeTotal;
115*4882a593Smuzhiyun struct eun_info_t {
116*4882a593Smuzhiyun uint32_t Offset;
117*4882a593Smuzhiyun uint32_t EraseCount;
118*4882a593Smuzhiyun uint32_t Free;
119*4882a593Smuzhiyun uint32_t Deleted;
120*4882a593Smuzhiyun } *EUNInfo;
121*4882a593Smuzhiyun struct xfer_info_t {
122*4882a593Smuzhiyun uint32_t Offset;
123*4882a593Smuzhiyun uint32_t EraseCount;
124*4882a593Smuzhiyun uint16_t state;
125*4882a593Smuzhiyun } *XferInfo;
126*4882a593Smuzhiyun uint16_t bam_index;
127*4882a593Smuzhiyun uint32_t *bam_cache;
128*4882a593Smuzhiyun uint16_t DataUnits;
129*4882a593Smuzhiyun uint32_t BlocksPerUnit;
130*4882a593Smuzhiyun erase_unit_header_t header;
131*4882a593Smuzhiyun } partition_t;
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun /* Partition state flags */
134*4882a593Smuzhiyun #define FTL_FORMATTED 0x01
135*4882a593Smuzhiyun
136*4882a593Smuzhiyun /* Transfer unit states */
137*4882a593Smuzhiyun #define XFER_UNKNOWN 0x00
138*4882a593Smuzhiyun #define XFER_ERASING 0x01
139*4882a593Smuzhiyun #define XFER_ERASED 0x02
140*4882a593Smuzhiyun #define XFER_PREPARED 0x03
141*4882a593Smuzhiyun #define XFER_FAILED 0x04
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun /*======================================================================
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun Scan_header() checks to see if a memory region contains an FTL
146*4882a593Smuzhiyun partition. build_maps() reads all the erase unit headers, builds
147*4882a593Smuzhiyun the erase unit map, and then builds the virtual page map.
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun ======================================================================*/
150*4882a593Smuzhiyun
scan_header(partition_t * part)151*4882a593Smuzhiyun static int scan_header(partition_t *part)
152*4882a593Smuzhiyun {
153*4882a593Smuzhiyun erase_unit_header_t header;
154*4882a593Smuzhiyun loff_t offset, max_offset;
155*4882a593Smuzhiyun size_t ret;
156*4882a593Smuzhiyun int err;
157*4882a593Smuzhiyun part->header.FormattedSize = 0;
158*4882a593Smuzhiyun max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size;
159*4882a593Smuzhiyun /* Search first megabyte for a valid FTL header */
160*4882a593Smuzhiyun for (offset = 0;
161*4882a593Smuzhiyun (offset + sizeof(header)) < max_offset;
162*4882a593Smuzhiyun offset += part->mbd.mtd->erasesize ? : 0x2000) {
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret,
165*4882a593Smuzhiyun (unsigned char *)&header);
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun if (err)
168*4882a593Smuzhiyun return err;
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun if (strcmp(header.DataOrgTuple+3, "FTL100") == 0) break;
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun if (offset == max_offset) {
174*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: FTL header not found.\n");
175*4882a593Smuzhiyun return -ENOENT;
176*4882a593Smuzhiyun }
177*4882a593Smuzhiyun if (header.BlockSize != 9 ||
178*4882a593Smuzhiyun (header.EraseUnitSize < 10) || (header.EraseUnitSize > 31) ||
179*4882a593Smuzhiyun (header.NumTransferUnits >= le16_to_cpu(header.NumEraseUnits))) {
180*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: FTL header corrupt!\n");
181*4882a593Smuzhiyun return -1;
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) {
184*4882a593Smuzhiyun printk(KERN_NOTICE "ftl: FTL EraseUnitSize %x != MTD erasesize %x\n",
185*4882a593Smuzhiyun 1 << header.EraseUnitSize,part->mbd.mtd->erasesize);
186*4882a593Smuzhiyun return -1;
187*4882a593Smuzhiyun }
188*4882a593Smuzhiyun part->header = header;
189*4882a593Smuzhiyun return 0;
190*4882a593Smuzhiyun }
191*4882a593Smuzhiyun
build_maps(partition_t * part)192*4882a593Smuzhiyun static int build_maps(partition_t *part)
193*4882a593Smuzhiyun {
194*4882a593Smuzhiyun erase_unit_header_t header;
195*4882a593Smuzhiyun uint16_t xvalid, xtrans, i;
196*4882a593Smuzhiyun unsigned blocks, j;
197*4882a593Smuzhiyun int hdr_ok, ret = -1;
198*4882a593Smuzhiyun ssize_t retval;
199*4882a593Smuzhiyun loff_t offset;
200*4882a593Smuzhiyun
201*4882a593Smuzhiyun /* Set up erase unit maps */
202*4882a593Smuzhiyun part->DataUnits = le16_to_cpu(part->header.NumEraseUnits) -
203*4882a593Smuzhiyun part->header.NumTransferUnits;
204*4882a593Smuzhiyun part->EUNInfo = kmalloc_array(part->DataUnits, sizeof(struct eun_info_t),
205*4882a593Smuzhiyun GFP_KERNEL);
206*4882a593Smuzhiyun if (!part->EUNInfo)
207*4882a593Smuzhiyun goto out;
208*4882a593Smuzhiyun for (i = 0; i < part->DataUnits; i++)
209*4882a593Smuzhiyun part->EUNInfo[i].Offset = 0xffffffff;
210*4882a593Smuzhiyun part->XferInfo =
211*4882a593Smuzhiyun kmalloc_array(part->header.NumTransferUnits,
212*4882a593Smuzhiyun sizeof(struct xfer_info_t),
213*4882a593Smuzhiyun GFP_KERNEL);
214*4882a593Smuzhiyun if (!part->XferInfo)
215*4882a593Smuzhiyun goto out_EUNInfo;
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun xvalid = xtrans = 0;
218*4882a593Smuzhiyun for (i = 0; i < le16_to_cpu(part->header.NumEraseUnits); i++) {
219*4882a593Smuzhiyun offset = ((i + le16_to_cpu(part->header.FirstPhysicalEUN))
220*4882a593Smuzhiyun << part->header.EraseUnitSize);
221*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, offset, sizeof(header), &retval,
222*4882a593Smuzhiyun (unsigned char *)&header);
223*4882a593Smuzhiyun
224*4882a593Smuzhiyun if (ret)
225*4882a593Smuzhiyun goto out_XferInfo;
226*4882a593Smuzhiyun
227*4882a593Smuzhiyun ret = -1;
228*4882a593Smuzhiyun /* Is this a transfer partition? */
229*4882a593Smuzhiyun hdr_ok = (strcmp(header.DataOrgTuple+3, "FTL100") == 0);
230*4882a593Smuzhiyun if (hdr_ok && (le16_to_cpu(header.LogicalEUN) < part->DataUnits) &&
231*4882a593Smuzhiyun (part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset == 0xffffffff)) {
232*4882a593Smuzhiyun part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset = offset;
233*4882a593Smuzhiyun part->EUNInfo[le16_to_cpu(header.LogicalEUN)].EraseCount =
234*4882a593Smuzhiyun le32_to_cpu(header.EraseCount);
235*4882a593Smuzhiyun xvalid++;
236*4882a593Smuzhiyun } else {
237*4882a593Smuzhiyun if (xtrans == part->header.NumTransferUnits) {
238*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: format error: too many "
239*4882a593Smuzhiyun "transfer units!\n");
240*4882a593Smuzhiyun goto out_XferInfo;
241*4882a593Smuzhiyun }
242*4882a593Smuzhiyun if (hdr_ok && (le16_to_cpu(header.LogicalEUN) == 0xffff)) {
243*4882a593Smuzhiyun part->XferInfo[xtrans].state = XFER_PREPARED;
244*4882a593Smuzhiyun part->XferInfo[xtrans].EraseCount = le32_to_cpu(header.EraseCount);
245*4882a593Smuzhiyun } else {
246*4882a593Smuzhiyun part->XferInfo[xtrans].state = XFER_UNKNOWN;
247*4882a593Smuzhiyun /* Pick anything reasonable for the erase count */
248*4882a593Smuzhiyun part->XferInfo[xtrans].EraseCount =
249*4882a593Smuzhiyun le32_to_cpu(part->header.EraseCount);
250*4882a593Smuzhiyun }
251*4882a593Smuzhiyun part->XferInfo[xtrans].Offset = offset;
252*4882a593Smuzhiyun xtrans++;
253*4882a593Smuzhiyun }
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun /* Check for format trouble */
256*4882a593Smuzhiyun header = part->header;
257*4882a593Smuzhiyun if ((xtrans != header.NumTransferUnits) ||
258*4882a593Smuzhiyun (xvalid+xtrans != le16_to_cpu(header.NumEraseUnits))) {
259*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: format error: erase units "
260*4882a593Smuzhiyun "don't add up!\n");
261*4882a593Smuzhiyun goto out_XferInfo;
262*4882a593Smuzhiyun }
263*4882a593Smuzhiyun
264*4882a593Smuzhiyun /* Set up virtual page map */
265*4882a593Smuzhiyun blocks = le32_to_cpu(header.FormattedSize) >> header.BlockSize;
266*4882a593Smuzhiyun part->VirtualBlockMap = vmalloc(array_size(blocks, sizeof(uint32_t)));
267*4882a593Smuzhiyun if (!part->VirtualBlockMap)
268*4882a593Smuzhiyun goto out_XferInfo;
269*4882a593Smuzhiyun
270*4882a593Smuzhiyun memset(part->VirtualBlockMap, 0xff, blocks * sizeof(uint32_t));
271*4882a593Smuzhiyun part->BlocksPerUnit = (1 << header.EraseUnitSize) >> header.BlockSize;
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun part->bam_cache = kmalloc_array(part->BlocksPerUnit, sizeof(uint32_t),
274*4882a593Smuzhiyun GFP_KERNEL);
275*4882a593Smuzhiyun if (!part->bam_cache)
276*4882a593Smuzhiyun goto out_VirtualBlockMap;
277*4882a593Smuzhiyun
278*4882a593Smuzhiyun part->bam_index = 0xffff;
279*4882a593Smuzhiyun part->FreeTotal = 0;
280*4882a593Smuzhiyun
281*4882a593Smuzhiyun for (i = 0; i < part->DataUnits; i++) {
282*4882a593Smuzhiyun part->EUNInfo[i].Free = 0;
283*4882a593Smuzhiyun part->EUNInfo[i].Deleted = 0;
284*4882a593Smuzhiyun offset = part->EUNInfo[i].Offset + le32_to_cpu(header.BAMOffset);
285*4882a593Smuzhiyun
286*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, offset,
287*4882a593Smuzhiyun part->BlocksPerUnit * sizeof(uint32_t), &retval,
288*4882a593Smuzhiyun (unsigned char *)part->bam_cache);
289*4882a593Smuzhiyun
290*4882a593Smuzhiyun if (ret)
291*4882a593Smuzhiyun goto out_bam_cache;
292*4882a593Smuzhiyun
293*4882a593Smuzhiyun for (j = 0; j < part->BlocksPerUnit; j++) {
294*4882a593Smuzhiyun if (BLOCK_FREE(le32_to_cpu(part->bam_cache[j]))) {
295*4882a593Smuzhiyun part->EUNInfo[i].Free++;
296*4882a593Smuzhiyun part->FreeTotal++;
297*4882a593Smuzhiyun } else if ((BLOCK_TYPE(le32_to_cpu(part->bam_cache[j])) == BLOCK_DATA) &&
298*4882a593Smuzhiyun (BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j])) < blocks))
299*4882a593Smuzhiyun part->VirtualBlockMap[BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j]))] =
300*4882a593Smuzhiyun (i << header.EraseUnitSize) + (j << header.BlockSize);
301*4882a593Smuzhiyun else if (BLOCK_DELETED(le32_to_cpu(part->bam_cache[j])))
302*4882a593Smuzhiyun part->EUNInfo[i].Deleted++;
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun }
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun ret = 0;
307*4882a593Smuzhiyun goto out;
308*4882a593Smuzhiyun
309*4882a593Smuzhiyun out_bam_cache:
310*4882a593Smuzhiyun kfree(part->bam_cache);
311*4882a593Smuzhiyun out_VirtualBlockMap:
312*4882a593Smuzhiyun vfree(part->VirtualBlockMap);
313*4882a593Smuzhiyun out_XferInfo:
314*4882a593Smuzhiyun kfree(part->XferInfo);
315*4882a593Smuzhiyun out_EUNInfo:
316*4882a593Smuzhiyun kfree(part->EUNInfo);
317*4882a593Smuzhiyun out:
318*4882a593Smuzhiyun return ret;
319*4882a593Smuzhiyun } /* build_maps */
320*4882a593Smuzhiyun
321*4882a593Smuzhiyun /*======================================================================
322*4882a593Smuzhiyun
323*4882a593Smuzhiyun Erase_xfer() schedules an asynchronous erase operation for a
324*4882a593Smuzhiyun transfer unit.
325*4882a593Smuzhiyun
326*4882a593Smuzhiyun ======================================================================*/
327*4882a593Smuzhiyun
erase_xfer(partition_t * part,uint16_t xfernum)328*4882a593Smuzhiyun static int erase_xfer(partition_t *part,
329*4882a593Smuzhiyun uint16_t xfernum)
330*4882a593Smuzhiyun {
331*4882a593Smuzhiyun int ret;
332*4882a593Smuzhiyun struct xfer_info_t *xfer;
333*4882a593Smuzhiyun struct erase_info *erase;
334*4882a593Smuzhiyun
335*4882a593Smuzhiyun xfer = &part->XferInfo[xfernum];
336*4882a593Smuzhiyun pr_debug("ftl_cs: erasing xfer unit at 0x%x\n", xfer->Offset);
337*4882a593Smuzhiyun xfer->state = XFER_ERASING;
338*4882a593Smuzhiyun
339*4882a593Smuzhiyun /* Is there a free erase slot? Always in MTD. */
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun
342*4882a593Smuzhiyun erase=kmalloc(sizeof(struct erase_info), GFP_KERNEL);
343*4882a593Smuzhiyun if (!erase)
344*4882a593Smuzhiyun return -ENOMEM;
345*4882a593Smuzhiyun
346*4882a593Smuzhiyun erase->addr = xfer->Offset;
347*4882a593Smuzhiyun erase->len = 1 << part->header.EraseUnitSize;
348*4882a593Smuzhiyun
349*4882a593Smuzhiyun ret = mtd_erase(part->mbd.mtd, erase);
350*4882a593Smuzhiyun if (!ret) {
351*4882a593Smuzhiyun xfer->state = XFER_ERASED;
352*4882a593Smuzhiyun xfer->EraseCount++;
353*4882a593Smuzhiyun } else {
354*4882a593Smuzhiyun xfer->state = XFER_FAILED;
355*4882a593Smuzhiyun pr_notice("ftl_cs: erase failed: err = %d\n", ret);
356*4882a593Smuzhiyun }
357*4882a593Smuzhiyun
358*4882a593Smuzhiyun kfree(erase);
359*4882a593Smuzhiyun
360*4882a593Smuzhiyun return ret;
361*4882a593Smuzhiyun } /* erase_xfer */
362*4882a593Smuzhiyun
363*4882a593Smuzhiyun /*======================================================================
364*4882a593Smuzhiyun
365*4882a593Smuzhiyun Prepare_xfer() takes a freshly erased transfer unit and gives
366*4882a593Smuzhiyun it an appropriate header.
367*4882a593Smuzhiyun
368*4882a593Smuzhiyun ======================================================================*/
369*4882a593Smuzhiyun
prepare_xfer(partition_t * part,int i)370*4882a593Smuzhiyun static int prepare_xfer(partition_t *part, int i)
371*4882a593Smuzhiyun {
372*4882a593Smuzhiyun erase_unit_header_t header;
373*4882a593Smuzhiyun struct xfer_info_t *xfer;
374*4882a593Smuzhiyun int nbam, ret;
375*4882a593Smuzhiyun uint32_t ctl;
376*4882a593Smuzhiyun ssize_t retlen;
377*4882a593Smuzhiyun loff_t offset;
378*4882a593Smuzhiyun
379*4882a593Smuzhiyun xfer = &part->XferInfo[i];
380*4882a593Smuzhiyun xfer->state = XFER_FAILED;
381*4882a593Smuzhiyun
382*4882a593Smuzhiyun pr_debug("ftl_cs: preparing xfer unit at 0x%x\n", xfer->Offset);
383*4882a593Smuzhiyun
384*4882a593Smuzhiyun /* Write the transfer unit header */
385*4882a593Smuzhiyun header = part->header;
386*4882a593Smuzhiyun header.LogicalEUN = cpu_to_le16(0xffff);
387*4882a593Smuzhiyun header.EraseCount = cpu_to_le32(xfer->EraseCount);
388*4882a593Smuzhiyun
389*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, xfer->Offset, sizeof(header), &retlen,
390*4882a593Smuzhiyun (u_char *)&header);
391*4882a593Smuzhiyun
392*4882a593Smuzhiyun if (ret) {
393*4882a593Smuzhiyun return ret;
394*4882a593Smuzhiyun }
395*4882a593Smuzhiyun
396*4882a593Smuzhiyun /* Write the BAM stub */
397*4882a593Smuzhiyun nbam = DIV_ROUND_UP(part->BlocksPerUnit * sizeof(uint32_t) +
398*4882a593Smuzhiyun le32_to_cpu(part->header.BAMOffset), SECTOR_SIZE);
399*4882a593Smuzhiyun
400*4882a593Smuzhiyun offset = xfer->Offset + le32_to_cpu(part->header.BAMOffset);
401*4882a593Smuzhiyun ctl = cpu_to_le32(BLOCK_CONTROL);
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun for (i = 0; i < nbam; i++, offset += sizeof(uint32_t)) {
404*4882a593Smuzhiyun
405*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
406*4882a593Smuzhiyun (u_char *)&ctl);
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun if (ret)
409*4882a593Smuzhiyun return ret;
410*4882a593Smuzhiyun }
411*4882a593Smuzhiyun xfer->state = XFER_PREPARED;
412*4882a593Smuzhiyun return 0;
413*4882a593Smuzhiyun
414*4882a593Smuzhiyun } /* prepare_xfer */
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun /*======================================================================
417*4882a593Smuzhiyun
418*4882a593Smuzhiyun Copy_erase_unit() takes a full erase block and a transfer unit,
419*4882a593Smuzhiyun copies everything to the transfer unit, then swaps the block
420*4882a593Smuzhiyun pointers.
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun All data blocks are copied to the corresponding blocks in the
423*4882a593Smuzhiyun target unit, so the virtual block map does not need to be
424*4882a593Smuzhiyun updated.
425*4882a593Smuzhiyun
426*4882a593Smuzhiyun ======================================================================*/
427*4882a593Smuzhiyun
copy_erase_unit(partition_t * part,uint16_t srcunit,uint16_t xferunit)428*4882a593Smuzhiyun static int copy_erase_unit(partition_t *part, uint16_t srcunit,
429*4882a593Smuzhiyun uint16_t xferunit)
430*4882a593Smuzhiyun {
431*4882a593Smuzhiyun u_char buf[SECTOR_SIZE];
432*4882a593Smuzhiyun struct eun_info_t *eun;
433*4882a593Smuzhiyun struct xfer_info_t *xfer;
434*4882a593Smuzhiyun uint32_t src, dest, free, i;
435*4882a593Smuzhiyun uint16_t unit;
436*4882a593Smuzhiyun int ret;
437*4882a593Smuzhiyun ssize_t retlen;
438*4882a593Smuzhiyun loff_t offset;
439*4882a593Smuzhiyun uint16_t srcunitswap = cpu_to_le16(srcunit);
440*4882a593Smuzhiyun
441*4882a593Smuzhiyun eun = &part->EUNInfo[srcunit];
442*4882a593Smuzhiyun xfer = &part->XferInfo[xferunit];
443*4882a593Smuzhiyun pr_debug("ftl_cs: copying block 0x%x to 0x%x\n",
444*4882a593Smuzhiyun eun->Offset, xfer->Offset);
445*4882a593Smuzhiyun
446*4882a593Smuzhiyun
447*4882a593Smuzhiyun /* Read current BAM */
448*4882a593Smuzhiyun if (part->bam_index != srcunit) {
449*4882a593Smuzhiyun
450*4882a593Smuzhiyun offset = eun->Offset + le32_to_cpu(part->header.BAMOffset);
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, offset,
453*4882a593Smuzhiyun part->BlocksPerUnit * sizeof(uint32_t), &retlen,
454*4882a593Smuzhiyun (u_char *)(part->bam_cache));
455*4882a593Smuzhiyun
456*4882a593Smuzhiyun /* mark the cache bad, in case we get an error later */
457*4882a593Smuzhiyun part->bam_index = 0xffff;
458*4882a593Smuzhiyun
459*4882a593Smuzhiyun if (ret) {
460*4882a593Smuzhiyun printk( KERN_WARNING "ftl: Failed to read BAM cache in copy_erase_unit()!\n");
461*4882a593Smuzhiyun return ret;
462*4882a593Smuzhiyun }
463*4882a593Smuzhiyun }
464*4882a593Smuzhiyun
465*4882a593Smuzhiyun /* Write the LogicalEUN for the transfer unit */
466*4882a593Smuzhiyun xfer->state = XFER_UNKNOWN;
467*4882a593Smuzhiyun offset = xfer->Offset + 20; /* Bad! */
468*4882a593Smuzhiyun unit = cpu_to_le16(0x7fff);
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, offset, sizeof(uint16_t), &retlen,
471*4882a593Smuzhiyun (u_char *)&unit);
472*4882a593Smuzhiyun
473*4882a593Smuzhiyun if (ret) {
474*4882a593Smuzhiyun printk( KERN_WARNING "ftl: Failed to write back to BAM cache in copy_erase_unit()!\n");
475*4882a593Smuzhiyun return ret;
476*4882a593Smuzhiyun }
477*4882a593Smuzhiyun
478*4882a593Smuzhiyun /* Copy all data blocks from source unit to transfer unit */
479*4882a593Smuzhiyun src = eun->Offset; dest = xfer->Offset;
480*4882a593Smuzhiyun
481*4882a593Smuzhiyun free = 0;
482*4882a593Smuzhiyun ret = 0;
483*4882a593Smuzhiyun for (i = 0; i < part->BlocksPerUnit; i++) {
484*4882a593Smuzhiyun switch (BLOCK_TYPE(le32_to_cpu(part->bam_cache[i]))) {
485*4882a593Smuzhiyun case BLOCK_CONTROL:
486*4882a593Smuzhiyun /* This gets updated later */
487*4882a593Smuzhiyun break;
488*4882a593Smuzhiyun case BLOCK_DATA:
489*4882a593Smuzhiyun case BLOCK_REPLACEMENT:
490*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, src, SECTOR_SIZE, &retlen,
491*4882a593Smuzhiyun (u_char *)buf);
492*4882a593Smuzhiyun if (ret) {
493*4882a593Smuzhiyun printk(KERN_WARNING "ftl: Error reading old xfer unit in copy_erase_unit\n");
494*4882a593Smuzhiyun return ret;
495*4882a593Smuzhiyun }
496*4882a593Smuzhiyun
497*4882a593Smuzhiyun
498*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, dest, SECTOR_SIZE, &retlen,
499*4882a593Smuzhiyun (u_char *)buf);
500*4882a593Smuzhiyun if (ret) {
501*4882a593Smuzhiyun printk(KERN_WARNING "ftl: Error writing new xfer unit in copy_erase_unit\n");
502*4882a593Smuzhiyun return ret;
503*4882a593Smuzhiyun }
504*4882a593Smuzhiyun
505*4882a593Smuzhiyun break;
506*4882a593Smuzhiyun default:
507*4882a593Smuzhiyun /* All other blocks must be free */
508*4882a593Smuzhiyun part->bam_cache[i] = cpu_to_le32(0xffffffff);
509*4882a593Smuzhiyun free++;
510*4882a593Smuzhiyun break;
511*4882a593Smuzhiyun }
512*4882a593Smuzhiyun src += SECTOR_SIZE;
513*4882a593Smuzhiyun dest += SECTOR_SIZE;
514*4882a593Smuzhiyun }
515*4882a593Smuzhiyun
516*4882a593Smuzhiyun /* Write the BAM to the transfer unit */
517*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd,
518*4882a593Smuzhiyun xfer->Offset + le32_to_cpu(part->header.BAMOffset),
519*4882a593Smuzhiyun part->BlocksPerUnit * sizeof(int32_t),
520*4882a593Smuzhiyun &retlen,
521*4882a593Smuzhiyun (u_char *)part->bam_cache);
522*4882a593Smuzhiyun if (ret) {
523*4882a593Smuzhiyun printk( KERN_WARNING "ftl: Error writing BAM in copy_erase_unit\n");
524*4882a593Smuzhiyun return ret;
525*4882a593Smuzhiyun }
526*4882a593Smuzhiyun
527*4882a593Smuzhiyun
528*4882a593Smuzhiyun /* All clear? Then update the LogicalEUN again */
529*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, xfer->Offset + 20, sizeof(uint16_t),
530*4882a593Smuzhiyun &retlen, (u_char *)&srcunitswap);
531*4882a593Smuzhiyun
532*4882a593Smuzhiyun if (ret) {
533*4882a593Smuzhiyun printk(KERN_WARNING "ftl: Error writing new LogicalEUN in copy_erase_unit\n");
534*4882a593Smuzhiyun return ret;
535*4882a593Smuzhiyun }
536*4882a593Smuzhiyun
537*4882a593Smuzhiyun
538*4882a593Smuzhiyun /* Update the maps and usage stats*/
539*4882a593Smuzhiyun swap(xfer->EraseCount, eun->EraseCount);
540*4882a593Smuzhiyun swap(xfer->Offset, eun->Offset);
541*4882a593Smuzhiyun part->FreeTotal -= eun->Free;
542*4882a593Smuzhiyun part->FreeTotal += free;
543*4882a593Smuzhiyun eun->Free = free;
544*4882a593Smuzhiyun eun->Deleted = 0;
545*4882a593Smuzhiyun
546*4882a593Smuzhiyun /* Now, the cache should be valid for the new block */
547*4882a593Smuzhiyun part->bam_index = srcunit;
548*4882a593Smuzhiyun
549*4882a593Smuzhiyun return 0;
550*4882a593Smuzhiyun } /* copy_erase_unit */
551*4882a593Smuzhiyun
552*4882a593Smuzhiyun /*======================================================================
553*4882a593Smuzhiyun
554*4882a593Smuzhiyun reclaim_block() picks a full erase unit and a transfer unit and
555*4882a593Smuzhiyun then calls copy_erase_unit() to copy one to the other. Then, it
556*4882a593Smuzhiyun schedules an erase on the expired block.
557*4882a593Smuzhiyun
558*4882a593Smuzhiyun What's a good way to decide which transfer unit and which erase
559*4882a593Smuzhiyun unit to use? Beats me. My way is to always pick the transfer
560*4882a593Smuzhiyun unit with the fewest erases, and usually pick the data unit with
561*4882a593Smuzhiyun the most deleted blocks. But with a small probability, pick the
562*4882a593Smuzhiyun oldest data unit instead. This means that we generally postpone
563*4882a593Smuzhiyun the next reclamation as long as possible, but shuffle static
564*4882a593Smuzhiyun stuff around a bit for wear leveling.
565*4882a593Smuzhiyun
566*4882a593Smuzhiyun ======================================================================*/
567*4882a593Smuzhiyun
reclaim_block(partition_t * part)568*4882a593Smuzhiyun static int reclaim_block(partition_t *part)
569*4882a593Smuzhiyun {
570*4882a593Smuzhiyun uint16_t i, eun, xfer;
571*4882a593Smuzhiyun uint32_t best;
572*4882a593Smuzhiyun int queued, ret;
573*4882a593Smuzhiyun
574*4882a593Smuzhiyun pr_debug("ftl_cs: reclaiming space...\n");
575*4882a593Smuzhiyun pr_debug("NumTransferUnits == %x\n", part->header.NumTransferUnits);
576*4882a593Smuzhiyun /* Pick the least erased transfer unit */
577*4882a593Smuzhiyun best = 0xffffffff; xfer = 0xffff;
578*4882a593Smuzhiyun do {
579*4882a593Smuzhiyun queued = 0;
580*4882a593Smuzhiyun for (i = 0; i < part->header.NumTransferUnits; i++) {
581*4882a593Smuzhiyun int n=0;
582*4882a593Smuzhiyun if (part->XferInfo[i].state == XFER_UNKNOWN) {
583*4882a593Smuzhiyun pr_debug("XferInfo[%d].state == XFER_UNKNOWN\n",i);
584*4882a593Smuzhiyun n=1;
585*4882a593Smuzhiyun erase_xfer(part, i);
586*4882a593Smuzhiyun }
587*4882a593Smuzhiyun if (part->XferInfo[i].state == XFER_ERASING) {
588*4882a593Smuzhiyun pr_debug("XferInfo[%d].state == XFER_ERASING\n",i);
589*4882a593Smuzhiyun n=1;
590*4882a593Smuzhiyun queued = 1;
591*4882a593Smuzhiyun }
592*4882a593Smuzhiyun else if (part->XferInfo[i].state == XFER_ERASED) {
593*4882a593Smuzhiyun pr_debug("XferInfo[%d].state == XFER_ERASED\n",i);
594*4882a593Smuzhiyun n=1;
595*4882a593Smuzhiyun prepare_xfer(part, i);
596*4882a593Smuzhiyun }
597*4882a593Smuzhiyun if (part->XferInfo[i].state == XFER_PREPARED) {
598*4882a593Smuzhiyun pr_debug("XferInfo[%d].state == XFER_PREPARED\n",i);
599*4882a593Smuzhiyun n=1;
600*4882a593Smuzhiyun if (part->XferInfo[i].EraseCount <= best) {
601*4882a593Smuzhiyun best = part->XferInfo[i].EraseCount;
602*4882a593Smuzhiyun xfer = i;
603*4882a593Smuzhiyun }
604*4882a593Smuzhiyun }
605*4882a593Smuzhiyun if (!n)
606*4882a593Smuzhiyun pr_debug("XferInfo[%d].state == %x\n",i, part->XferInfo[i].state);
607*4882a593Smuzhiyun
608*4882a593Smuzhiyun }
609*4882a593Smuzhiyun if (xfer == 0xffff) {
610*4882a593Smuzhiyun if (queued) {
611*4882a593Smuzhiyun pr_debug("ftl_cs: waiting for transfer "
612*4882a593Smuzhiyun "unit to be prepared...\n");
613*4882a593Smuzhiyun mtd_sync(part->mbd.mtd);
614*4882a593Smuzhiyun } else {
615*4882a593Smuzhiyun static int ne = 0;
616*4882a593Smuzhiyun if (++ne < 5)
617*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: reclaim failed: no "
618*4882a593Smuzhiyun "suitable transfer units!\n");
619*4882a593Smuzhiyun else
620*4882a593Smuzhiyun pr_debug("ftl_cs: reclaim failed: no "
621*4882a593Smuzhiyun "suitable transfer units!\n");
622*4882a593Smuzhiyun
623*4882a593Smuzhiyun return -EIO;
624*4882a593Smuzhiyun }
625*4882a593Smuzhiyun }
626*4882a593Smuzhiyun } while (xfer == 0xffff);
627*4882a593Smuzhiyun
628*4882a593Smuzhiyun eun = 0;
629*4882a593Smuzhiyun if ((jiffies % shuffle_freq) == 0) {
630*4882a593Smuzhiyun pr_debug("ftl_cs: recycling freshest block...\n");
631*4882a593Smuzhiyun best = 0xffffffff;
632*4882a593Smuzhiyun for (i = 0; i < part->DataUnits; i++)
633*4882a593Smuzhiyun if (part->EUNInfo[i].EraseCount <= best) {
634*4882a593Smuzhiyun best = part->EUNInfo[i].EraseCount;
635*4882a593Smuzhiyun eun = i;
636*4882a593Smuzhiyun }
637*4882a593Smuzhiyun } else {
638*4882a593Smuzhiyun best = 0;
639*4882a593Smuzhiyun for (i = 0; i < part->DataUnits; i++)
640*4882a593Smuzhiyun if (part->EUNInfo[i].Deleted >= best) {
641*4882a593Smuzhiyun best = part->EUNInfo[i].Deleted;
642*4882a593Smuzhiyun eun = i;
643*4882a593Smuzhiyun }
644*4882a593Smuzhiyun if (best == 0) {
645*4882a593Smuzhiyun static int ne = 0;
646*4882a593Smuzhiyun if (++ne < 5)
647*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: reclaim failed: "
648*4882a593Smuzhiyun "no free blocks!\n");
649*4882a593Smuzhiyun else
650*4882a593Smuzhiyun pr_debug("ftl_cs: reclaim failed: "
651*4882a593Smuzhiyun "no free blocks!\n");
652*4882a593Smuzhiyun
653*4882a593Smuzhiyun return -EIO;
654*4882a593Smuzhiyun }
655*4882a593Smuzhiyun }
656*4882a593Smuzhiyun ret = copy_erase_unit(part, eun, xfer);
657*4882a593Smuzhiyun if (!ret)
658*4882a593Smuzhiyun erase_xfer(part, xfer);
659*4882a593Smuzhiyun else
660*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: copy_erase_unit failed!\n");
661*4882a593Smuzhiyun return ret;
662*4882a593Smuzhiyun } /* reclaim_block */
663*4882a593Smuzhiyun
664*4882a593Smuzhiyun /*======================================================================
665*4882a593Smuzhiyun
666*4882a593Smuzhiyun Find_free() searches for a free block. If necessary, it updates
667*4882a593Smuzhiyun the BAM cache for the erase unit containing the free block. It
668*4882a593Smuzhiyun returns the block index -- the erase unit is just the currently
669*4882a593Smuzhiyun cached unit. If there are no free blocks, it returns 0 -- this
670*4882a593Smuzhiyun is never a valid data block because it contains the header.
671*4882a593Smuzhiyun
672*4882a593Smuzhiyun ======================================================================*/
673*4882a593Smuzhiyun
674*4882a593Smuzhiyun #ifdef PSYCHO_DEBUG
dump_lists(partition_t * part)675*4882a593Smuzhiyun static void dump_lists(partition_t *part)
676*4882a593Smuzhiyun {
677*4882a593Smuzhiyun int i;
678*4882a593Smuzhiyun printk(KERN_DEBUG "ftl_cs: Free total = %d\n", part->FreeTotal);
679*4882a593Smuzhiyun for (i = 0; i < part->DataUnits; i++)
680*4882a593Smuzhiyun printk(KERN_DEBUG "ftl_cs: unit %d: %d phys, %d free, "
681*4882a593Smuzhiyun "%d deleted\n", i,
682*4882a593Smuzhiyun part->EUNInfo[i].Offset >> part->header.EraseUnitSize,
683*4882a593Smuzhiyun part->EUNInfo[i].Free, part->EUNInfo[i].Deleted);
684*4882a593Smuzhiyun }
685*4882a593Smuzhiyun #endif
686*4882a593Smuzhiyun
find_free(partition_t * part)687*4882a593Smuzhiyun static uint32_t find_free(partition_t *part)
688*4882a593Smuzhiyun {
689*4882a593Smuzhiyun uint16_t stop, eun;
690*4882a593Smuzhiyun uint32_t blk;
691*4882a593Smuzhiyun size_t retlen;
692*4882a593Smuzhiyun int ret;
693*4882a593Smuzhiyun
694*4882a593Smuzhiyun /* Find an erase unit with some free space */
695*4882a593Smuzhiyun stop = (part->bam_index == 0xffff) ? 0 : part->bam_index;
696*4882a593Smuzhiyun eun = stop;
697*4882a593Smuzhiyun do {
698*4882a593Smuzhiyun if (part->EUNInfo[eun].Free != 0) break;
699*4882a593Smuzhiyun /* Wrap around at end of table */
700*4882a593Smuzhiyun if (++eun == part->DataUnits) eun = 0;
701*4882a593Smuzhiyun } while (eun != stop);
702*4882a593Smuzhiyun
703*4882a593Smuzhiyun if (part->EUNInfo[eun].Free == 0)
704*4882a593Smuzhiyun return 0;
705*4882a593Smuzhiyun
706*4882a593Smuzhiyun /* Is this unit's BAM cached? */
707*4882a593Smuzhiyun if (eun != part->bam_index) {
708*4882a593Smuzhiyun /* Invalidate cache */
709*4882a593Smuzhiyun part->bam_index = 0xffff;
710*4882a593Smuzhiyun
711*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd,
712*4882a593Smuzhiyun part->EUNInfo[eun].Offset + le32_to_cpu(part->header.BAMOffset),
713*4882a593Smuzhiyun part->BlocksPerUnit * sizeof(uint32_t),
714*4882a593Smuzhiyun &retlen,
715*4882a593Smuzhiyun (u_char *)(part->bam_cache));
716*4882a593Smuzhiyun
717*4882a593Smuzhiyun if (ret) {
718*4882a593Smuzhiyun printk(KERN_WARNING"ftl: Error reading BAM in find_free\n");
719*4882a593Smuzhiyun return 0;
720*4882a593Smuzhiyun }
721*4882a593Smuzhiyun part->bam_index = eun;
722*4882a593Smuzhiyun }
723*4882a593Smuzhiyun
724*4882a593Smuzhiyun /* Find a free block */
725*4882a593Smuzhiyun for (blk = 0; blk < part->BlocksPerUnit; blk++)
726*4882a593Smuzhiyun if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break;
727*4882a593Smuzhiyun if (blk == part->BlocksPerUnit) {
728*4882a593Smuzhiyun #ifdef PSYCHO_DEBUG
729*4882a593Smuzhiyun static int ne = 0;
730*4882a593Smuzhiyun if (++ne == 1)
731*4882a593Smuzhiyun dump_lists(part);
732*4882a593Smuzhiyun #endif
733*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: bad free list!\n");
734*4882a593Smuzhiyun return 0;
735*4882a593Smuzhiyun }
736*4882a593Smuzhiyun pr_debug("ftl_cs: found free block at %d in %d\n", blk, eun);
737*4882a593Smuzhiyun return blk;
738*4882a593Smuzhiyun
739*4882a593Smuzhiyun } /* find_free */
740*4882a593Smuzhiyun
741*4882a593Smuzhiyun
742*4882a593Smuzhiyun /*======================================================================
743*4882a593Smuzhiyun
744*4882a593Smuzhiyun Read a series of sectors from an FTL partition.
745*4882a593Smuzhiyun
746*4882a593Smuzhiyun ======================================================================*/
747*4882a593Smuzhiyun
ftl_read(partition_t * part,caddr_t buffer,u_long sector,u_long nblocks)748*4882a593Smuzhiyun static int ftl_read(partition_t *part, caddr_t buffer,
749*4882a593Smuzhiyun u_long sector, u_long nblocks)
750*4882a593Smuzhiyun {
751*4882a593Smuzhiyun uint32_t log_addr, bsize;
752*4882a593Smuzhiyun u_long i;
753*4882a593Smuzhiyun int ret;
754*4882a593Smuzhiyun size_t offset, retlen;
755*4882a593Smuzhiyun
756*4882a593Smuzhiyun pr_debug("ftl_cs: ftl_read(0x%p, 0x%lx, %ld)\n",
757*4882a593Smuzhiyun part, sector, nblocks);
758*4882a593Smuzhiyun if (!(part->state & FTL_FORMATTED)) {
759*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: bad partition\n");
760*4882a593Smuzhiyun return -EIO;
761*4882a593Smuzhiyun }
762*4882a593Smuzhiyun bsize = 1 << part->header.EraseUnitSize;
763*4882a593Smuzhiyun
764*4882a593Smuzhiyun for (i = 0; i < nblocks; i++) {
765*4882a593Smuzhiyun if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) {
766*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: bad read offset\n");
767*4882a593Smuzhiyun return -EIO;
768*4882a593Smuzhiyun }
769*4882a593Smuzhiyun log_addr = part->VirtualBlockMap[sector+i];
770*4882a593Smuzhiyun if (log_addr == 0xffffffff)
771*4882a593Smuzhiyun memset(buffer, 0, SECTOR_SIZE);
772*4882a593Smuzhiyun else {
773*4882a593Smuzhiyun offset = (part->EUNInfo[log_addr / bsize].Offset
774*4882a593Smuzhiyun + (log_addr % bsize));
775*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, offset, SECTOR_SIZE, &retlen,
776*4882a593Smuzhiyun (u_char *)buffer);
777*4882a593Smuzhiyun
778*4882a593Smuzhiyun if (ret) {
779*4882a593Smuzhiyun printk(KERN_WARNING "Error reading MTD device in ftl_read()\n");
780*4882a593Smuzhiyun return ret;
781*4882a593Smuzhiyun }
782*4882a593Smuzhiyun }
783*4882a593Smuzhiyun buffer += SECTOR_SIZE;
784*4882a593Smuzhiyun }
785*4882a593Smuzhiyun return 0;
786*4882a593Smuzhiyun } /* ftl_read */
787*4882a593Smuzhiyun
788*4882a593Smuzhiyun /*======================================================================
789*4882a593Smuzhiyun
790*4882a593Smuzhiyun Write a series of sectors to an FTL partition
791*4882a593Smuzhiyun
792*4882a593Smuzhiyun ======================================================================*/
793*4882a593Smuzhiyun
set_bam_entry(partition_t * part,uint32_t log_addr,uint32_t virt_addr)794*4882a593Smuzhiyun static int set_bam_entry(partition_t *part, uint32_t log_addr,
795*4882a593Smuzhiyun uint32_t virt_addr)
796*4882a593Smuzhiyun {
797*4882a593Smuzhiyun uint32_t bsize, blk, le_virt_addr;
798*4882a593Smuzhiyun #ifdef PSYCHO_DEBUG
799*4882a593Smuzhiyun uint32_t old_addr;
800*4882a593Smuzhiyun #endif
801*4882a593Smuzhiyun uint16_t eun;
802*4882a593Smuzhiyun int ret;
803*4882a593Smuzhiyun size_t retlen, offset;
804*4882a593Smuzhiyun
805*4882a593Smuzhiyun pr_debug("ftl_cs: set_bam_entry(0x%p, 0x%x, 0x%x)\n",
806*4882a593Smuzhiyun part, log_addr, virt_addr);
807*4882a593Smuzhiyun bsize = 1 << part->header.EraseUnitSize;
808*4882a593Smuzhiyun eun = log_addr / bsize;
809*4882a593Smuzhiyun blk = (log_addr % bsize) / SECTOR_SIZE;
810*4882a593Smuzhiyun offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) +
811*4882a593Smuzhiyun le32_to_cpu(part->header.BAMOffset));
812*4882a593Smuzhiyun
813*4882a593Smuzhiyun #ifdef PSYCHO_DEBUG
814*4882a593Smuzhiyun ret = mtd_read(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
815*4882a593Smuzhiyun (u_char *)&old_addr);
816*4882a593Smuzhiyun if (ret) {
817*4882a593Smuzhiyun printk(KERN_WARNING"ftl: Error reading old_addr in set_bam_entry: %d\n",ret);
818*4882a593Smuzhiyun return ret;
819*4882a593Smuzhiyun }
820*4882a593Smuzhiyun old_addr = le32_to_cpu(old_addr);
821*4882a593Smuzhiyun
822*4882a593Smuzhiyun if (((virt_addr == 0xfffffffe) && !BLOCK_FREE(old_addr)) ||
823*4882a593Smuzhiyun ((virt_addr == 0) && (BLOCK_TYPE(old_addr) != BLOCK_DATA)) ||
824*4882a593Smuzhiyun (!BLOCK_DELETED(virt_addr) && (old_addr != 0xfffffffe))) {
825*4882a593Smuzhiyun static int ne = 0;
826*4882a593Smuzhiyun if (++ne < 5) {
827*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: set_bam_entry() inconsistency!\n");
828*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: log_addr = 0x%x, old = 0x%x"
829*4882a593Smuzhiyun ", new = 0x%x\n", log_addr, old_addr, virt_addr);
830*4882a593Smuzhiyun }
831*4882a593Smuzhiyun return -EIO;
832*4882a593Smuzhiyun }
833*4882a593Smuzhiyun #endif
834*4882a593Smuzhiyun le_virt_addr = cpu_to_le32(virt_addr);
835*4882a593Smuzhiyun if (part->bam_index == eun) {
836*4882a593Smuzhiyun #ifdef PSYCHO_DEBUG
837*4882a593Smuzhiyun if (le32_to_cpu(part->bam_cache[blk]) != old_addr) {
838*4882a593Smuzhiyun static int ne = 0;
839*4882a593Smuzhiyun if (++ne < 5) {
840*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: set_bam_entry() "
841*4882a593Smuzhiyun "inconsistency!\n");
842*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: log_addr = 0x%x, cache"
843*4882a593Smuzhiyun " = 0x%x\n",
844*4882a593Smuzhiyun le32_to_cpu(part->bam_cache[blk]), old_addr);
845*4882a593Smuzhiyun }
846*4882a593Smuzhiyun return -EIO;
847*4882a593Smuzhiyun }
848*4882a593Smuzhiyun #endif
849*4882a593Smuzhiyun part->bam_cache[blk] = le_virt_addr;
850*4882a593Smuzhiyun }
851*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
852*4882a593Smuzhiyun (u_char *)&le_virt_addr);
853*4882a593Smuzhiyun
854*4882a593Smuzhiyun if (ret) {
855*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: set_bam_entry() failed!\n");
856*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: log_addr = 0x%x, new = 0x%x\n",
857*4882a593Smuzhiyun log_addr, virt_addr);
858*4882a593Smuzhiyun }
859*4882a593Smuzhiyun return ret;
860*4882a593Smuzhiyun } /* set_bam_entry */
861*4882a593Smuzhiyun
ftl_write(partition_t * part,caddr_t buffer,u_long sector,u_long nblocks)862*4882a593Smuzhiyun static int ftl_write(partition_t *part, caddr_t buffer,
863*4882a593Smuzhiyun u_long sector, u_long nblocks)
864*4882a593Smuzhiyun {
865*4882a593Smuzhiyun uint32_t bsize, log_addr, virt_addr, old_addr, blk;
866*4882a593Smuzhiyun u_long i;
867*4882a593Smuzhiyun int ret;
868*4882a593Smuzhiyun size_t retlen, offset;
869*4882a593Smuzhiyun
870*4882a593Smuzhiyun pr_debug("ftl_cs: ftl_write(0x%p, %ld, %ld)\n",
871*4882a593Smuzhiyun part, sector, nblocks);
872*4882a593Smuzhiyun if (!(part->state & FTL_FORMATTED)) {
873*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: bad partition\n");
874*4882a593Smuzhiyun return -EIO;
875*4882a593Smuzhiyun }
876*4882a593Smuzhiyun /* See if we need to reclaim space, before we start */
877*4882a593Smuzhiyun while (part->FreeTotal < nblocks) {
878*4882a593Smuzhiyun ret = reclaim_block(part);
879*4882a593Smuzhiyun if (ret)
880*4882a593Smuzhiyun return ret;
881*4882a593Smuzhiyun }
882*4882a593Smuzhiyun
883*4882a593Smuzhiyun bsize = 1 << part->header.EraseUnitSize;
884*4882a593Smuzhiyun
885*4882a593Smuzhiyun virt_addr = sector * SECTOR_SIZE | BLOCK_DATA;
886*4882a593Smuzhiyun for (i = 0; i < nblocks; i++) {
887*4882a593Smuzhiyun if (virt_addr >= le32_to_cpu(part->header.FormattedSize)) {
888*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: bad write offset\n");
889*4882a593Smuzhiyun return -EIO;
890*4882a593Smuzhiyun }
891*4882a593Smuzhiyun
892*4882a593Smuzhiyun /* Grab a free block */
893*4882a593Smuzhiyun blk = find_free(part);
894*4882a593Smuzhiyun if (blk == 0) {
895*4882a593Smuzhiyun static int ne = 0;
896*4882a593Smuzhiyun if (++ne < 5)
897*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: internal error: "
898*4882a593Smuzhiyun "no free blocks!\n");
899*4882a593Smuzhiyun return -ENOSPC;
900*4882a593Smuzhiyun }
901*4882a593Smuzhiyun
902*4882a593Smuzhiyun /* Tag the BAM entry, and write the new block */
903*4882a593Smuzhiyun log_addr = part->bam_index * bsize + blk * SECTOR_SIZE;
904*4882a593Smuzhiyun part->EUNInfo[part->bam_index].Free--;
905*4882a593Smuzhiyun part->FreeTotal--;
906*4882a593Smuzhiyun if (set_bam_entry(part, log_addr, 0xfffffffe))
907*4882a593Smuzhiyun return -EIO;
908*4882a593Smuzhiyun part->EUNInfo[part->bam_index].Deleted++;
909*4882a593Smuzhiyun offset = (part->EUNInfo[part->bam_index].Offset +
910*4882a593Smuzhiyun blk * SECTOR_SIZE);
911*4882a593Smuzhiyun ret = mtd_write(part->mbd.mtd, offset, SECTOR_SIZE, &retlen, buffer);
912*4882a593Smuzhiyun
913*4882a593Smuzhiyun if (ret) {
914*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: block write failed!\n");
915*4882a593Smuzhiyun printk(KERN_NOTICE "ftl_cs: log_addr = 0x%x, virt_addr"
916*4882a593Smuzhiyun " = 0x%x, Offset = 0x%zx\n", log_addr, virt_addr,
917*4882a593Smuzhiyun offset);
918*4882a593Smuzhiyun return -EIO;
919*4882a593Smuzhiyun }
920*4882a593Smuzhiyun
921*4882a593Smuzhiyun /* Only delete the old entry when the new entry is ready */
922*4882a593Smuzhiyun old_addr = part->VirtualBlockMap[sector+i];
923*4882a593Smuzhiyun if (old_addr != 0xffffffff) {
924*4882a593Smuzhiyun part->VirtualBlockMap[sector+i] = 0xffffffff;
925*4882a593Smuzhiyun part->EUNInfo[old_addr/bsize].Deleted++;
926*4882a593Smuzhiyun if (set_bam_entry(part, old_addr, 0))
927*4882a593Smuzhiyun return -EIO;
928*4882a593Smuzhiyun }
929*4882a593Smuzhiyun
930*4882a593Smuzhiyun /* Finally, set up the new pointers */
931*4882a593Smuzhiyun if (set_bam_entry(part, log_addr, virt_addr))
932*4882a593Smuzhiyun return -EIO;
933*4882a593Smuzhiyun part->VirtualBlockMap[sector+i] = log_addr;
934*4882a593Smuzhiyun part->EUNInfo[part->bam_index].Deleted--;
935*4882a593Smuzhiyun
936*4882a593Smuzhiyun buffer += SECTOR_SIZE;
937*4882a593Smuzhiyun virt_addr += SECTOR_SIZE;
938*4882a593Smuzhiyun }
939*4882a593Smuzhiyun return 0;
940*4882a593Smuzhiyun } /* ftl_write */
941*4882a593Smuzhiyun
ftl_getgeo(struct mtd_blktrans_dev * dev,struct hd_geometry * geo)942*4882a593Smuzhiyun static int ftl_getgeo(struct mtd_blktrans_dev *dev, struct hd_geometry *geo)
943*4882a593Smuzhiyun {
944*4882a593Smuzhiyun partition_t *part = (void *)dev;
945*4882a593Smuzhiyun u_long sect;
946*4882a593Smuzhiyun
947*4882a593Smuzhiyun /* Sort of arbitrary: round size down to 4KiB boundary */
948*4882a593Smuzhiyun sect = le32_to_cpu(part->header.FormattedSize)/SECTOR_SIZE;
949*4882a593Smuzhiyun
950*4882a593Smuzhiyun geo->heads = 1;
951*4882a593Smuzhiyun geo->sectors = 8;
952*4882a593Smuzhiyun geo->cylinders = sect >> 3;
953*4882a593Smuzhiyun
954*4882a593Smuzhiyun return 0;
955*4882a593Smuzhiyun }
956*4882a593Smuzhiyun
ftl_readsect(struct mtd_blktrans_dev * dev,unsigned long block,char * buf)957*4882a593Smuzhiyun static int ftl_readsect(struct mtd_blktrans_dev *dev,
958*4882a593Smuzhiyun unsigned long block, char *buf)
959*4882a593Smuzhiyun {
960*4882a593Smuzhiyun return ftl_read((void *)dev, buf, block, 1);
961*4882a593Smuzhiyun }
962*4882a593Smuzhiyun
ftl_writesect(struct mtd_blktrans_dev * dev,unsigned long block,char * buf)963*4882a593Smuzhiyun static int ftl_writesect(struct mtd_blktrans_dev *dev,
964*4882a593Smuzhiyun unsigned long block, char *buf)
965*4882a593Smuzhiyun {
966*4882a593Smuzhiyun return ftl_write((void *)dev, buf, block, 1);
967*4882a593Smuzhiyun }
968*4882a593Smuzhiyun
ftl_discardsect(struct mtd_blktrans_dev * dev,unsigned long sector,unsigned nr_sects)969*4882a593Smuzhiyun static int ftl_discardsect(struct mtd_blktrans_dev *dev,
970*4882a593Smuzhiyun unsigned long sector, unsigned nr_sects)
971*4882a593Smuzhiyun {
972*4882a593Smuzhiyun partition_t *part = (void *)dev;
973*4882a593Smuzhiyun uint32_t bsize = 1 << part->header.EraseUnitSize;
974*4882a593Smuzhiyun
975*4882a593Smuzhiyun pr_debug("FTL erase sector %ld for %d sectors\n",
976*4882a593Smuzhiyun sector, nr_sects);
977*4882a593Smuzhiyun
978*4882a593Smuzhiyun while (nr_sects) {
979*4882a593Smuzhiyun uint32_t old_addr = part->VirtualBlockMap[sector];
980*4882a593Smuzhiyun if (old_addr != 0xffffffff) {
981*4882a593Smuzhiyun part->VirtualBlockMap[sector] = 0xffffffff;
982*4882a593Smuzhiyun part->EUNInfo[old_addr/bsize].Deleted++;
983*4882a593Smuzhiyun if (set_bam_entry(part, old_addr, 0))
984*4882a593Smuzhiyun return -EIO;
985*4882a593Smuzhiyun }
986*4882a593Smuzhiyun nr_sects--;
987*4882a593Smuzhiyun sector++;
988*4882a593Smuzhiyun }
989*4882a593Smuzhiyun
990*4882a593Smuzhiyun return 0;
991*4882a593Smuzhiyun }
992*4882a593Smuzhiyun /*====================================================================*/
993*4882a593Smuzhiyun
ftl_freepart(partition_t * part)994*4882a593Smuzhiyun static void ftl_freepart(partition_t *part)
995*4882a593Smuzhiyun {
996*4882a593Smuzhiyun vfree(part->VirtualBlockMap);
997*4882a593Smuzhiyun part->VirtualBlockMap = NULL;
998*4882a593Smuzhiyun kfree(part->EUNInfo);
999*4882a593Smuzhiyun part->EUNInfo = NULL;
1000*4882a593Smuzhiyun kfree(part->XferInfo);
1001*4882a593Smuzhiyun part->XferInfo = NULL;
1002*4882a593Smuzhiyun kfree(part->bam_cache);
1003*4882a593Smuzhiyun part->bam_cache = NULL;
1004*4882a593Smuzhiyun } /* ftl_freepart */
1005*4882a593Smuzhiyun
ftl_add_mtd(struct mtd_blktrans_ops * tr,struct mtd_info * mtd)1006*4882a593Smuzhiyun static void ftl_add_mtd(struct mtd_blktrans_ops *tr, struct mtd_info *mtd)
1007*4882a593Smuzhiyun {
1008*4882a593Smuzhiyun partition_t *partition;
1009*4882a593Smuzhiyun
1010*4882a593Smuzhiyun partition = kzalloc(sizeof(partition_t), GFP_KERNEL);
1011*4882a593Smuzhiyun
1012*4882a593Smuzhiyun if (!partition) {
1013*4882a593Smuzhiyun printk(KERN_WARNING "No memory to scan for FTL on %s\n",
1014*4882a593Smuzhiyun mtd->name);
1015*4882a593Smuzhiyun return;
1016*4882a593Smuzhiyun }
1017*4882a593Smuzhiyun
1018*4882a593Smuzhiyun partition->mbd.mtd = mtd;
1019*4882a593Smuzhiyun
1020*4882a593Smuzhiyun if ((scan_header(partition) == 0) &&
1021*4882a593Smuzhiyun (build_maps(partition) == 0)) {
1022*4882a593Smuzhiyun
1023*4882a593Smuzhiyun partition->state = FTL_FORMATTED;
1024*4882a593Smuzhiyun #ifdef PCMCIA_DEBUG
1025*4882a593Smuzhiyun printk(KERN_INFO "ftl_cs: opening %d KiB FTL partition\n",
1026*4882a593Smuzhiyun le32_to_cpu(partition->header.FormattedSize) >> 10);
1027*4882a593Smuzhiyun #endif
1028*4882a593Smuzhiyun partition->mbd.size = le32_to_cpu(partition->header.FormattedSize) >> 9;
1029*4882a593Smuzhiyun
1030*4882a593Smuzhiyun partition->mbd.tr = tr;
1031*4882a593Smuzhiyun partition->mbd.devnum = -1;
1032*4882a593Smuzhiyun if (!add_mtd_blktrans_dev((void *)partition))
1033*4882a593Smuzhiyun return;
1034*4882a593Smuzhiyun }
1035*4882a593Smuzhiyun
1036*4882a593Smuzhiyun kfree(partition);
1037*4882a593Smuzhiyun }
1038*4882a593Smuzhiyun
ftl_remove_dev(struct mtd_blktrans_dev * dev)1039*4882a593Smuzhiyun static void ftl_remove_dev(struct mtd_blktrans_dev *dev)
1040*4882a593Smuzhiyun {
1041*4882a593Smuzhiyun del_mtd_blktrans_dev(dev);
1042*4882a593Smuzhiyun ftl_freepart((partition_t *)dev);
1043*4882a593Smuzhiyun }
1044*4882a593Smuzhiyun
1045*4882a593Smuzhiyun static struct mtd_blktrans_ops ftl_tr = {
1046*4882a593Smuzhiyun .name = "ftl",
1047*4882a593Smuzhiyun .major = FTL_MAJOR,
1048*4882a593Smuzhiyun .part_bits = PART_BITS,
1049*4882a593Smuzhiyun .blksize = SECTOR_SIZE,
1050*4882a593Smuzhiyun .readsect = ftl_readsect,
1051*4882a593Smuzhiyun .writesect = ftl_writesect,
1052*4882a593Smuzhiyun .discard = ftl_discardsect,
1053*4882a593Smuzhiyun .getgeo = ftl_getgeo,
1054*4882a593Smuzhiyun .add_mtd = ftl_add_mtd,
1055*4882a593Smuzhiyun .remove_dev = ftl_remove_dev,
1056*4882a593Smuzhiyun .owner = THIS_MODULE,
1057*4882a593Smuzhiyun };
1058*4882a593Smuzhiyun
init_ftl(void)1059*4882a593Smuzhiyun static int __init init_ftl(void)
1060*4882a593Smuzhiyun {
1061*4882a593Smuzhiyun return register_mtd_blktrans(&ftl_tr);
1062*4882a593Smuzhiyun }
1063*4882a593Smuzhiyun
cleanup_ftl(void)1064*4882a593Smuzhiyun static void __exit cleanup_ftl(void)
1065*4882a593Smuzhiyun {
1066*4882a593Smuzhiyun deregister_mtd_blktrans(&ftl_tr);
1067*4882a593Smuzhiyun }
1068*4882a593Smuzhiyun
1069*4882a593Smuzhiyun module_init(init_ftl);
1070*4882a593Smuzhiyun module_exit(cleanup_ftl);
1071*4882a593Smuzhiyun
1072*4882a593Smuzhiyun
1073*4882a593Smuzhiyun MODULE_LICENSE("Dual MPL/GPL");
1074*4882a593Smuzhiyun MODULE_AUTHOR("David Hinds <dahinds@users.sourceforge.net>");
1075*4882a593Smuzhiyun MODULE_DESCRIPTION("Support code for Flash Translation Layer, used on PCMCIA devices");
1076