dev_mtd.c 6.1 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253
  1. /*
  2. * fs/logfs/dev_mtd.c - Device access methods for MTD
  3. *
  4. * As should be obvious for Linux kernel code, license is GPLv2
  5. *
  6. * Copyright (c) 2005-2008 Joern Engel <joern@logfs.org>
  7. */
  8. #include "logfs.h"
  9. #include <linux/completion.h>
  10. #include <linux/mount.h>
  11. #include <linux/sched.h>
  12. #define PAGE_OFS(ofs) ((ofs) & (PAGE_SIZE-1))
  13. static int mtd_read(struct super_block *sb, loff_t ofs, size_t len, void *buf)
  14. {
  15. struct mtd_info *mtd = logfs_super(sb)->s_mtd;
  16. size_t retlen;
  17. int ret;
  18. ret = mtd->read(mtd, ofs, len, &retlen, buf);
  19. BUG_ON(ret == -EINVAL);
  20. if (ret)
  21. return ret;
  22. /* Not sure if we should loop instead. */
  23. if (retlen != len)
  24. return -EIO;
  25. return 0;
  26. }
  27. static int mtd_write(struct super_block *sb, loff_t ofs, size_t len, void *buf)
  28. {
  29. struct logfs_super *super = logfs_super(sb);
  30. struct mtd_info *mtd = super->s_mtd;
  31. size_t retlen;
  32. loff_t page_start, page_end;
  33. int ret;
  34. if (super->s_flags & LOGFS_SB_FLAG_RO)
  35. return -EROFS;
  36. BUG_ON((ofs >= mtd->size) || (len > mtd->size - ofs));
  37. BUG_ON(ofs != (ofs >> super->s_writeshift) << super->s_writeshift);
  38. BUG_ON(len > PAGE_CACHE_SIZE);
  39. page_start = ofs & PAGE_CACHE_MASK;
  40. page_end = PAGE_CACHE_ALIGN(ofs + len) - 1;
  41. ret = mtd->write(mtd, ofs, len, &retlen, buf);
  42. if (ret || (retlen != len))
  43. return -EIO;
  44. return 0;
  45. }
  46. /*
  47. * For as long as I can remember (since about 2001) mtd->erase has been an
  48. * asynchronous interface lacking the first driver to actually use the
  49. * asynchronous properties. So just to prevent the first implementor of such
  50. * a thing from breaking logfs in 2350, we do the usual pointless dance to
  51. * declare a completion variable and wait for completion before returning
  52. * from mtd_erase(). What an excercise in futility!
  53. */
  54. static void logfs_erase_callback(struct erase_info *ei)
  55. {
  56. complete((struct completion *)ei->priv);
  57. }
  58. static int mtd_erase_mapping(struct super_block *sb, loff_t ofs, size_t len)
  59. {
  60. struct logfs_super *super = logfs_super(sb);
  61. struct address_space *mapping = super->s_mapping_inode->i_mapping;
  62. struct page *page;
  63. pgoff_t index = ofs >> PAGE_SHIFT;
  64. for (index = ofs >> PAGE_SHIFT; index < (ofs + len) >> PAGE_SHIFT; index++) {
  65. page = find_get_page(mapping, index);
  66. if (!page)
  67. continue;
  68. memset(page_address(page), 0xFF, PAGE_SIZE);
  69. page_cache_release(page);
  70. }
  71. return 0;
  72. }
  73. static int mtd_erase(struct super_block *sb, loff_t ofs, size_t len)
  74. {
  75. struct mtd_info *mtd = logfs_super(sb)->s_mtd;
  76. struct erase_info ei;
  77. DECLARE_COMPLETION_ONSTACK(complete);
  78. int ret;
  79. BUG_ON(len % mtd->erasesize);
  80. if (logfs_super(sb)->s_flags & LOGFS_SB_FLAG_RO)
  81. return -EROFS;
  82. memset(&ei, 0, sizeof(ei));
  83. ei.mtd = mtd;
  84. ei.addr = ofs;
  85. ei.len = len;
  86. ei.callback = logfs_erase_callback;
  87. ei.priv = (long)&complete;
  88. ret = mtd->erase(mtd, &ei);
  89. if (ret)
  90. return -EIO;
  91. wait_for_completion(&complete);
  92. if (ei.state != MTD_ERASE_DONE)
  93. return -EIO;
  94. return mtd_erase_mapping(sb, ofs, len);
  95. }
  96. static void mtd_sync(struct super_block *sb)
  97. {
  98. struct mtd_info *mtd = logfs_super(sb)->s_mtd;
  99. if (mtd->sync)
  100. mtd->sync(mtd);
  101. }
  102. static int mtd_readpage(void *_sb, struct page *page)
  103. {
  104. struct super_block *sb = _sb;
  105. int err;
  106. err = mtd_read(sb, page->index << PAGE_SHIFT, PAGE_SIZE,
  107. page_address(page));
  108. if (err == -EUCLEAN) {
  109. err = 0;
  110. /* FIXME: force GC this segment */
  111. }
  112. if (err) {
  113. ClearPageUptodate(page);
  114. SetPageError(page);
  115. } else {
  116. SetPageUptodate(page);
  117. ClearPageError(page);
  118. }
  119. unlock_page(page);
  120. return err;
  121. }
  122. static struct page *mtd_find_first_sb(struct super_block *sb, u64 *ofs)
  123. {
  124. struct logfs_super *super = logfs_super(sb);
  125. struct address_space *mapping = super->s_mapping_inode->i_mapping;
  126. filler_t *filler = mtd_readpage;
  127. struct mtd_info *mtd = super->s_mtd;
  128. if (!mtd->block_isbad)
  129. return NULL;
  130. *ofs = 0;
  131. while (mtd->block_isbad(mtd, *ofs)) {
  132. *ofs += mtd->erasesize;
  133. if (*ofs >= mtd->size)
  134. return NULL;
  135. }
  136. BUG_ON(*ofs & ~PAGE_MASK);
  137. return read_cache_page(mapping, *ofs >> PAGE_SHIFT, filler, sb);
  138. }
  139. static struct page *mtd_find_last_sb(struct super_block *sb, u64 *ofs)
  140. {
  141. struct logfs_super *super = logfs_super(sb);
  142. struct address_space *mapping = super->s_mapping_inode->i_mapping;
  143. filler_t *filler = mtd_readpage;
  144. struct mtd_info *mtd = super->s_mtd;
  145. if (!mtd->block_isbad)
  146. return NULL;
  147. *ofs = mtd->size - mtd->erasesize;
  148. while (mtd->block_isbad(mtd, *ofs)) {
  149. *ofs -= mtd->erasesize;
  150. if (*ofs <= 0)
  151. return NULL;
  152. }
  153. *ofs = *ofs + mtd->erasesize - 0x1000;
  154. BUG_ON(*ofs & ~PAGE_MASK);
  155. return read_cache_page(mapping, *ofs >> PAGE_SHIFT, filler, sb);
  156. }
  157. static int __mtd_writeseg(struct super_block *sb, u64 ofs, pgoff_t index,
  158. size_t nr_pages)
  159. {
  160. struct logfs_super *super = logfs_super(sb);
  161. struct address_space *mapping = super->s_mapping_inode->i_mapping;
  162. struct page *page;
  163. int i, err;
  164. for (i = 0; i < nr_pages; i++) {
  165. page = find_lock_page(mapping, index + i);
  166. BUG_ON(!page);
  167. err = mtd_write(sb, page->index << PAGE_SHIFT, PAGE_SIZE,
  168. page_address(page));
  169. unlock_page(page);
  170. page_cache_release(page);
  171. if (err)
  172. return err;
  173. }
  174. return 0;
  175. }
  176. static void mtd_writeseg(struct super_block *sb, u64 ofs, size_t len)
  177. {
  178. struct logfs_super *super = logfs_super(sb);
  179. int head;
  180. if (super->s_flags & LOGFS_SB_FLAG_RO)
  181. return;
  182. if (len == 0) {
  183. /* This can happen when the object fit perfectly into a
  184. * segment, the segment gets written per sync and subsequently
  185. * closed.
  186. */
  187. return;
  188. }
  189. head = ofs & (PAGE_SIZE - 1);
  190. if (head) {
  191. ofs -= head;
  192. len += head;
  193. }
  194. len = PAGE_ALIGN(len);
  195. __mtd_writeseg(sb, ofs, ofs >> PAGE_SHIFT, len >> PAGE_SHIFT);
  196. }
  197. static void mtd_put_device(struct super_block *sb)
  198. {
  199. put_mtd_device(logfs_super(sb)->s_mtd);
  200. }
  201. static const struct logfs_device_ops mtd_devops = {
  202. .find_first_sb = mtd_find_first_sb,
  203. .find_last_sb = mtd_find_last_sb,
  204. .readpage = mtd_readpage,
  205. .writeseg = mtd_writeseg,
  206. .erase = mtd_erase,
  207. .sync = mtd_sync,
  208. .put_device = mtd_put_device,
  209. };
  210. int logfs_get_sb_mtd(struct file_system_type *type, int flags,
  211. int mtdnr, struct vfsmount *mnt)
  212. {
  213. struct mtd_info *mtd;
  214. const struct logfs_device_ops *devops = &mtd_devops;
  215. mtd = get_mtd_device(NULL, mtdnr);
  216. return logfs_get_sb_device(type, flags, mtd, NULL, devops, mnt);
  217. }