drbd_main.c 98 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309231023112312231323142315231623172318231923202321232223232324232523262327232823292330233123322333233423352336233723382339234023412342234323442345234623472348234923502351235223532354235523562357235823592360236123622363236423652366236723682369237023712372237323742375237623772378237923802381238223832384238523862387238823892390239123922393239423952396239723982399240024012402240324042405240624072408240924102411241224132414241524162417241824192420242124222423242424252426242724282429243024312432243324342435243624372438243924402441244224432444244524462447244824492450245124522453245424552456245724582459246024612462246324642465246624672468246924702471247224732474247524762477247824792480248124822483248424852486248724882489249024912492249324942495249624972498249925002501250225032504250525062507250825092510251125122513251425152516251725182519252025212522252325242525252625272528252925302531253225332534253525362537253825392540254125422543254425452546254725482549255025512552255325542555255625572558255925602561256225632564256525662567256825692570257125722573257425752576257725782579258025812582258325842585258625872588258925902591259225932594259525962597259825992600260126022603260426052606260726082609261026112612261326142615261626172618261926202621262226232624262526262627262826292630263126322633263426352636263726382639264026412642264326442645264626472648264926502651265226532654265526562657265826592660266126622663266426652666266726682669267026712672267326742675267626772678267926802681268226832684268526862687268826892690269126922693269426952696269726982699270027012702270327042705270627072708270927102711271227132714271527162717271827192720272127222723272427252726272727282729273027312732273327342735273627372738273927402741274227432744274527462747274827492750275127522753275427552756275727582759276027612762276327642765276627672768276927702771277227732774277527762777277827792780278127822783278427852786278727882789279027912792279327942795279627972798279928002801280228032804280528062807280828092810281128122813281428152816281728182819282028212822282328242825282628272828282928302831283228332834283528362837283828392840284128422843284428452846284728482849285028512852285328542855285628572858285928602861286228632864286528662867286828692870287128722873287428752876287728782879288028812882288328842885288628872888288928902891289228932894289528962897289828992900290129022903290429052906290729082909291029112912291329142915291629172918291929202921292229232924292529262927292829292930293129322933293429352936293729382939294029412942294329442945294629472948294929502951295229532954295529562957295829592960296129622963296429652966296729682969297029712972297329742975297629772978297929802981298229832984298529862987298829892990299129922993299429952996299729982999300030013002300330043005300630073008300930103011301230133014301530163017301830193020302130223023302430253026302730283029303030313032303330343035303630373038303930403041304230433044304530463047304830493050305130523053305430553056305730583059306030613062306330643065306630673068306930703071307230733074307530763077307830793080308130823083308430853086308730883089309030913092309330943095309630973098309931003101310231033104310531063107310831093110311131123113311431153116311731183119312031213122312331243125312631273128312931303131313231333134313531363137313831393140314131423143314431453146314731483149315031513152315331543155315631573158315931603161316231633164316531663167316831693170317131723173317431753176317731783179318031813182318331843185318631873188318931903191319231933194319531963197319831993200320132023203320432053206320732083209321032113212321332143215321632173218321932203221322232233224322532263227322832293230323132323233323432353236323732383239324032413242324332443245324632473248324932503251325232533254325532563257325832593260326132623263326432653266326732683269327032713272327332743275327632773278327932803281328232833284328532863287328832893290329132923293329432953296329732983299330033013302330333043305330633073308330933103311331233133314331533163317331833193320332133223323332433253326332733283329333033313332333333343335333633373338333933403341334233433344334533463347334833493350335133523353335433553356335733583359336033613362336333643365336633673368336933703371337233733374337533763377337833793380338133823383338433853386338733883389339033913392339333943395339633973398339934003401340234033404340534063407340834093410341134123413341434153416341734183419342034213422342334243425342634273428342934303431343234333434343534363437343834393440344134423443344434453446344734483449345034513452345334543455345634573458345934603461346234633464346534663467346834693470347134723473347434753476347734783479348034813482348334843485348634873488348934903491349234933494349534963497349834993500350135023503350435053506350735083509351035113512351335143515351635173518351935203521352235233524352535263527352835293530353135323533353435353536353735383539354035413542354335443545354635473548354935503551355235533554355535563557355835593560356135623563356435653566356735683569357035713572357335743575357635773578357935803581358235833584358535863587358835893590359135923593359435953596359735983599360036013602360336043605360636073608360936103611361236133614361536163617361836193620362136223623362436253626362736283629363036313632363336343635363636373638363936403641364236433644364536463647364836493650365136523653365436553656365736583659366036613662366336643665366636673668366936703671367236733674
  1. /*
  2. drbd.c
  3. This file is part of DRBD by Philipp Reisner and Lars Ellenberg.
  4. Copyright (C) 2001-2008, LINBIT Information Technologies GmbH.
  5. Copyright (C) 1999-2008, Philipp Reisner <philipp.reisner@linbit.com>.
  6. Copyright (C) 2002-2008, Lars Ellenberg <lars.ellenberg@linbit.com>.
  7. Thanks to Carter Burden, Bart Grantham and Gennadiy Nerubayev
  8. from Logicworks, Inc. for making SDP replication support possible.
  9. drbd is free software; you can redistribute it and/or modify
  10. it under the terms of the GNU General Public License as published by
  11. the Free Software Foundation; either version 2, or (at your option)
  12. any later version.
  13. drbd is distributed in the hope that it will be useful,
  14. but WITHOUT ANY WARRANTY; without even the implied warranty of
  15. MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  16. GNU General Public License for more details.
  17. You should have received a copy of the GNU General Public License
  18. along with drbd; see the file COPYING. If not, write to
  19. the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
  20. */
  21. #include <linux/module.h>
  22. #include <linux/drbd.h>
  23. #include <asm/uaccess.h>
  24. #include <asm/types.h>
  25. #include <net/sock.h>
  26. #include <linux/ctype.h>
  27. #include <linux/mutex.h>
  28. #include <linux/fs.h>
  29. #include <linux/file.h>
  30. #include <linux/proc_fs.h>
  31. #include <linux/init.h>
  32. #include <linux/mm.h>
  33. #include <linux/memcontrol.h>
  34. #include <linux/mm_inline.h>
  35. #include <linux/slab.h>
  36. #include <linux/random.h>
  37. #include <linux/reboot.h>
  38. #include <linux/notifier.h>
  39. #include <linux/kthread.h>
  40. #define __KERNEL_SYSCALLS__
  41. #include <linux/unistd.h>
  42. #include <linux/vmalloc.h>
  43. #include <linux/drbd_limits.h>
  44. #include "drbd_int.h"
  45. #include "drbd_req.h" /* only for _req_mod in tl_release and tl_clear */
  46. #include "drbd_vli.h"
  47. static DEFINE_MUTEX(drbd_main_mutex);
  48. int drbdd_init(struct drbd_thread *);
  49. int drbd_worker(struct drbd_thread *);
  50. int drbd_asender(struct drbd_thread *);
  51. int drbd_init(void);
  52. static int drbd_open(struct block_device *bdev, fmode_t mode);
  53. static int drbd_release(struct gendisk *gd, fmode_t mode);
  54. static int w_md_sync(struct drbd_work *w, int unused);
  55. static void md_sync_timer_fn(unsigned long data);
  56. static int w_bitmap_io(struct drbd_work *w, int unused);
  57. static int w_go_diskless(struct drbd_work *w, int unused);
  58. MODULE_AUTHOR("Philipp Reisner <phil@linbit.com>, "
  59. "Lars Ellenberg <lars@linbit.com>");
  60. MODULE_DESCRIPTION("drbd - Distributed Replicated Block Device v" REL_VERSION);
  61. MODULE_VERSION(REL_VERSION);
  62. MODULE_LICENSE("GPL");
  63. MODULE_PARM_DESC(minor_count, "Approximate number of drbd devices ("
  64. __stringify(DRBD_MINOR_COUNT_MIN) "-" __stringify(DRBD_MINOR_COUNT_MAX) ")");
  65. MODULE_ALIAS_BLOCKDEV_MAJOR(DRBD_MAJOR);
  66. #include <linux/moduleparam.h>
  67. /* allow_open_on_secondary */
  68. MODULE_PARM_DESC(allow_oos, "DONT USE!");
  69. /* thanks to these macros, if compiled into the kernel (not-module),
  70. * this becomes the boot parameter drbd.minor_count */
  71. module_param(minor_count, uint, 0444);
  72. module_param(disable_sendpage, bool, 0644);
  73. module_param(allow_oos, bool, 0);
  74. module_param(proc_details, int, 0644);
  75. #ifdef CONFIG_DRBD_FAULT_INJECTION
  76. int enable_faults;
  77. int fault_rate;
  78. static int fault_count;
  79. int fault_devs;
  80. /* bitmap of enabled faults */
  81. module_param(enable_faults, int, 0664);
  82. /* fault rate % value - applies to all enabled faults */
  83. module_param(fault_rate, int, 0664);
  84. /* count of faults inserted */
  85. module_param(fault_count, int, 0664);
  86. /* bitmap of devices to insert faults on */
  87. module_param(fault_devs, int, 0644);
  88. #endif
  89. /* module parameter, defined */
  90. unsigned int minor_count = DRBD_MINOR_COUNT_DEF;
  91. int disable_sendpage;
  92. int allow_oos;
  93. int proc_details; /* Detail level in proc drbd*/
  94. /* Module parameter for setting the user mode helper program
  95. * to run. Default is /sbin/drbdadm */
  96. char usermode_helper[80] = "/sbin/drbdadm";
  97. module_param_string(usermode_helper, usermode_helper, sizeof(usermode_helper), 0644);
  98. /* in 2.6.x, our device mapping and config info contains our virtual gendisks
  99. * as member "struct gendisk *vdisk;"
  100. */
  101. struct idr minors;
  102. struct list_head drbd_tconns; /* list of struct drbd_tconn */
  103. struct kmem_cache *drbd_request_cache;
  104. struct kmem_cache *drbd_ee_cache; /* peer requests */
  105. struct kmem_cache *drbd_bm_ext_cache; /* bitmap extents */
  106. struct kmem_cache *drbd_al_ext_cache; /* activity log extents */
  107. mempool_t *drbd_request_mempool;
  108. mempool_t *drbd_ee_mempool;
  109. mempool_t *drbd_md_io_page_pool;
  110. struct bio_set *drbd_md_io_bio_set;
  111. /* I do not use a standard mempool, because:
  112. 1) I want to hand out the pre-allocated objects first.
  113. 2) I want to be able to interrupt sleeping allocation with a signal.
  114. Note: This is a single linked list, the next pointer is the private
  115. member of struct page.
  116. */
  117. struct page *drbd_pp_pool;
  118. spinlock_t drbd_pp_lock;
  119. int drbd_pp_vacant;
  120. wait_queue_head_t drbd_pp_wait;
  121. DEFINE_RATELIMIT_STATE(drbd_ratelimit_state, 5 * HZ, 5);
  122. static const struct block_device_operations drbd_ops = {
  123. .owner = THIS_MODULE,
  124. .open = drbd_open,
  125. .release = drbd_release,
  126. };
  127. static void bio_destructor_drbd(struct bio *bio)
  128. {
  129. bio_free(bio, drbd_md_io_bio_set);
  130. }
  131. struct bio *bio_alloc_drbd(gfp_t gfp_mask)
  132. {
  133. struct bio *bio;
  134. if (!drbd_md_io_bio_set)
  135. return bio_alloc(gfp_mask, 1);
  136. bio = bio_alloc_bioset(gfp_mask, 1, drbd_md_io_bio_set);
  137. if (!bio)
  138. return NULL;
  139. bio->bi_destructor = bio_destructor_drbd;
  140. return bio;
  141. }
  142. #ifdef __CHECKER__
  143. /* When checking with sparse, and this is an inline function, sparse will
  144. give tons of false positives. When this is a real functions sparse works.
  145. */
  146. int _get_ldev_if_state(struct drbd_conf *mdev, enum drbd_disk_state mins)
  147. {
  148. int io_allowed;
  149. atomic_inc(&mdev->local_cnt);
  150. io_allowed = (mdev->state.disk >= mins);
  151. if (!io_allowed) {
  152. if (atomic_dec_and_test(&mdev->local_cnt))
  153. wake_up(&mdev->misc_wait);
  154. }
  155. return io_allowed;
  156. }
  157. #endif
  158. /**
  159. * DOC: The transfer log
  160. *
  161. * The transfer log is a single linked list of &struct drbd_tl_epoch objects.
  162. * mdev->tconn->newest_tle points to the head, mdev->tconn->oldest_tle points to the tail
  163. * of the list. There is always at least one &struct drbd_tl_epoch object.
  164. *
  165. * Each &struct drbd_tl_epoch has a circular double linked list of requests
  166. * attached.
  167. */
  168. static int tl_init(struct drbd_tconn *tconn)
  169. {
  170. struct drbd_tl_epoch *b;
  171. /* during device minor initialization, we may well use GFP_KERNEL */
  172. b = kmalloc(sizeof(struct drbd_tl_epoch), GFP_KERNEL);
  173. if (!b)
  174. return 0;
  175. INIT_LIST_HEAD(&b->requests);
  176. INIT_LIST_HEAD(&b->w.list);
  177. b->next = NULL;
  178. b->br_number = 4711;
  179. b->n_writes = 0;
  180. b->w.cb = NULL; /* if this is != NULL, we need to dec_ap_pending in tl_clear */
  181. tconn->oldest_tle = b;
  182. tconn->newest_tle = b;
  183. INIT_LIST_HEAD(&tconn->out_of_sequence_requests);
  184. INIT_LIST_HEAD(&tconn->barrier_acked_requests);
  185. return 1;
  186. }
  187. static void tl_cleanup(struct drbd_tconn *tconn)
  188. {
  189. if (tconn->oldest_tle != tconn->newest_tle)
  190. conn_err(tconn, "ASSERT FAILED: oldest_tle == newest_tle\n");
  191. if (!list_empty(&tconn->out_of_sequence_requests))
  192. conn_err(tconn, "ASSERT FAILED: list_empty(out_of_sequence_requests)\n");
  193. kfree(tconn->oldest_tle);
  194. tconn->oldest_tle = NULL;
  195. kfree(tconn->unused_spare_tle);
  196. tconn->unused_spare_tle = NULL;
  197. }
  198. /**
  199. * _tl_add_barrier() - Adds a barrier to the transfer log
  200. * @mdev: DRBD device.
  201. * @new: Barrier to be added before the current head of the TL.
  202. *
  203. * The caller must hold the req_lock.
  204. */
  205. void _tl_add_barrier(struct drbd_tconn *tconn, struct drbd_tl_epoch *new)
  206. {
  207. struct drbd_tl_epoch *newest_before;
  208. INIT_LIST_HEAD(&new->requests);
  209. INIT_LIST_HEAD(&new->w.list);
  210. new->w.cb = NULL; /* if this is != NULL, we need to dec_ap_pending in tl_clear */
  211. new->next = NULL;
  212. new->n_writes = 0;
  213. newest_before = tconn->newest_tle;
  214. new->br_number = newest_before->br_number+1;
  215. if (tconn->newest_tle != new) {
  216. tconn->newest_tle->next = new;
  217. tconn->newest_tle = new;
  218. }
  219. }
  220. /**
  221. * tl_release() - Free or recycle the oldest &struct drbd_tl_epoch object of the TL
  222. * @mdev: DRBD device.
  223. * @barrier_nr: Expected identifier of the DRBD write barrier packet.
  224. * @set_size: Expected number of requests before that barrier.
  225. *
  226. * In case the passed barrier_nr or set_size does not match the oldest
  227. * &struct drbd_tl_epoch objects this function will cause a termination
  228. * of the connection.
  229. */
  230. void tl_release(struct drbd_tconn *tconn, unsigned int barrier_nr,
  231. unsigned int set_size)
  232. {
  233. struct drbd_conf *mdev;
  234. struct drbd_tl_epoch *b, *nob; /* next old barrier */
  235. struct list_head *le, *tle;
  236. struct drbd_request *r;
  237. spin_lock_irq(&tconn->req_lock);
  238. b = tconn->oldest_tle;
  239. /* first some paranoia code */
  240. if (b == NULL) {
  241. conn_err(tconn, "BAD! BarrierAck #%u received, but no epoch in tl!?\n",
  242. barrier_nr);
  243. goto bail;
  244. }
  245. if (b->br_number != barrier_nr) {
  246. conn_err(tconn, "BAD! BarrierAck #%u received, expected #%u!\n",
  247. barrier_nr, b->br_number);
  248. goto bail;
  249. }
  250. if (b->n_writes != set_size) {
  251. conn_err(tconn, "BAD! BarrierAck #%u received with n_writes=%u, expected n_writes=%u!\n",
  252. barrier_nr, set_size, b->n_writes);
  253. goto bail;
  254. }
  255. /* Clean up list of requests processed during current epoch */
  256. list_for_each_safe(le, tle, &b->requests) {
  257. r = list_entry(le, struct drbd_request, tl_requests);
  258. _req_mod(r, BARRIER_ACKED);
  259. }
  260. /* There could be requests on the list waiting for completion
  261. of the write to the local disk. To avoid corruptions of
  262. slab's data structures we have to remove the lists head.
  263. Also there could have been a barrier ack out of sequence, overtaking
  264. the write acks - which would be a bug and violating write ordering.
  265. To not deadlock in case we lose connection while such requests are
  266. still pending, we need some way to find them for the
  267. _req_mode(CONNECTION_LOST_WHILE_PENDING).
  268. These have been list_move'd to the out_of_sequence_requests list in
  269. _req_mod(, BARRIER_ACKED) above.
  270. */
  271. list_splice_init(&b->requests, &tconn->barrier_acked_requests);
  272. mdev = b->w.mdev;
  273. nob = b->next;
  274. if (test_and_clear_bit(CREATE_BARRIER, &tconn->flags)) {
  275. _tl_add_barrier(tconn, b);
  276. if (nob)
  277. tconn->oldest_tle = nob;
  278. /* if nob == NULL b was the only barrier, and becomes the new
  279. barrier. Therefore tconn->oldest_tle points already to b */
  280. } else {
  281. D_ASSERT(nob != NULL);
  282. tconn->oldest_tle = nob;
  283. kfree(b);
  284. }
  285. spin_unlock_irq(&tconn->req_lock);
  286. dec_ap_pending(mdev);
  287. return;
  288. bail:
  289. spin_unlock_irq(&tconn->req_lock);
  290. conn_request_state(tconn, NS(conn, C_PROTOCOL_ERROR), CS_HARD);
  291. }
  292. /**
  293. * _tl_restart() - Walks the transfer log, and applies an action to all requests
  294. * @mdev: DRBD device.
  295. * @what: The action/event to perform with all request objects
  296. *
  297. * @what might be one of CONNECTION_LOST_WHILE_PENDING, RESEND, FAIL_FROZEN_DISK_IO,
  298. * RESTART_FROZEN_DISK_IO.
  299. */
  300. void _tl_restart(struct drbd_tconn *tconn, enum drbd_req_event what)
  301. {
  302. struct drbd_tl_epoch *b, *tmp, **pn;
  303. struct list_head *le, *tle, carry_reads;
  304. struct drbd_request *req;
  305. int rv, n_writes, n_reads;
  306. b = tconn->oldest_tle;
  307. pn = &tconn->oldest_tle;
  308. while (b) {
  309. n_writes = 0;
  310. n_reads = 0;
  311. INIT_LIST_HEAD(&carry_reads);
  312. list_for_each_safe(le, tle, &b->requests) {
  313. req = list_entry(le, struct drbd_request, tl_requests);
  314. rv = _req_mod(req, what);
  315. if (rv & MR_WRITE)
  316. n_writes++;
  317. if (rv & MR_READ)
  318. n_reads++;
  319. }
  320. tmp = b->next;
  321. if (n_writes) {
  322. if (what == RESEND) {
  323. b->n_writes = n_writes;
  324. if (b->w.cb == NULL) {
  325. b->w.cb = w_send_barrier;
  326. inc_ap_pending(b->w.mdev);
  327. set_bit(CREATE_BARRIER, &tconn->flags);
  328. }
  329. drbd_queue_work(&tconn->data.work, &b->w);
  330. }
  331. pn = &b->next;
  332. } else {
  333. if (n_reads)
  334. list_add(&carry_reads, &b->requests);
  335. /* there could still be requests on that ring list,
  336. * in case local io is still pending */
  337. list_del(&b->requests);
  338. /* dec_ap_pending corresponding to queue_barrier.
  339. * the newest barrier may not have been queued yet,
  340. * in which case w.cb is still NULL. */
  341. if (b->w.cb != NULL)
  342. dec_ap_pending(b->w.mdev);
  343. if (b == tconn->newest_tle) {
  344. /* recycle, but reinit! */
  345. if (tmp != NULL)
  346. conn_err(tconn, "ASSERT FAILED tmp == NULL");
  347. INIT_LIST_HEAD(&b->requests);
  348. list_splice(&carry_reads, &b->requests);
  349. INIT_LIST_HEAD(&b->w.list);
  350. b->w.cb = NULL;
  351. b->br_number = net_random();
  352. b->n_writes = 0;
  353. *pn = b;
  354. break;
  355. }
  356. *pn = tmp;
  357. kfree(b);
  358. }
  359. b = tmp;
  360. list_splice(&carry_reads, &b->requests);
  361. }
  362. /* Actions operating on the disk state, also want to work on
  363. requests that got barrier acked. */
  364. switch (what) {
  365. case FAIL_FROZEN_DISK_IO:
  366. case RESTART_FROZEN_DISK_IO:
  367. list_for_each_safe(le, tle, &tconn->barrier_acked_requests) {
  368. req = list_entry(le, struct drbd_request, tl_requests);
  369. _req_mod(req, what);
  370. }
  371. case CONNECTION_LOST_WHILE_PENDING:
  372. case RESEND:
  373. break;
  374. default:
  375. conn_err(tconn, "what = %d in _tl_restart()\n", what);
  376. }
  377. }
  378. /**
  379. * tl_clear() - Clears all requests and &struct drbd_tl_epoch objects out of the TL
  380. * @mdev: DRBD device.
  381. *
  382. * This is called after the connection to the peer was lost. The storage covered
  383. * by the requests on the transfer gets marked as our of sync. Called from the
  384. * receiver thread and the worker thread.
  385. */
  386. void tl_clear(struct drbd_tconn *tconn)
  387. {
  388. struct list_head *le, *tle;
  389. struct drbd_request *r;
  390. spin_lock_irq(&tconn->req_lock);
  391. _tl_restart(tconn, CONNECTION_LOST_WHILE_PENDING);
  392. /* we expect this list to be empty. */
  393. if (!list_empty(&tconn->out_of_sequence_requests))
  394. conn_err(tconn, "ASSERT FAILED list_empty(&out_of_sequence_requests)\n");
  395. /* but just in case, clean it up anyways! */
  396. list_for_each_safe(le, tle, &tconn->out_of_sequence_requests) {
  397. r = list_entry(le, struct drbd_request, tl_requests);
  398. /* It would be nice to complete outside of spinlock.
  399. * But this is easier for now. */
  400. _req_mod(r, CONNECTION_LOST_WHILE_PENDING);
  401. }
  402. /* ensure bit indicating barrier is required is clear */
  403. clear_bit(CREATE_BARRIER, &tconn->flags);
  404. spin_unlock_irq(&tconn->req_lock);
  405. }
  406. void tl_restart(struct drbd_tconn *tconn, enum drbd_req_event what)
  407. {
  408. spin_lock_irq(&tconn->req_lock);
  409. _tl_restart(tconn, what);
  410. spin_unlock_irq(&tconn->req_lock);
  411. }
  412. /**
  413. * tl_abort_disk_io() - Abort disk I/O for all requests for a certain mdev in the TL
  414. * @mdev: DRBD device.
  415. */
  416. void tl_abort_disk_io(struct drbd_conf *mdev)
  417. {
  418. struct drbd_tconn *tconn = mdev->tconn;
  419. struct drbd_tl_epoch *b;
  420. struct list_head *le, *tle;
  421. struct drbd_request *req;
  422. spin_lock_irq(&tconn->req_lock);
  423. b = tconn->oldest_tle;
  424. while (b) {
  425. list_for_each_safe(le, tle, &b->requests) {
  426. req = list_entry(le, struct drbd_request, tl_requests);
  427. if (!(req->rq_state & RQ_LOCAL_PENDING))
  428. continue;
  429. if (req->w.mdev == mdev)
  430. _req_mod(req, ABORT_DISK_IO);
  431. }
  432. b = b->next;
  433. }
  434. list_for_each_safe(le, tle, &tconn->barrier_acked_requests) {
  435. req = list_entry(le, struct drbd_request, tl_requests);
  436. if (!(req->rq_state & RQ_LOCAL_PENDING))
  437. continue;
  438. if (req->w.mdev == mdev)
  439. _req_mod(req, ABORT_DISK_IO);
  440. }
  441. spin_unlock_irq(&tconn->req_lock);
  442. }
  443. static int drbd_thread_setup(void *arg)
  444. {
  445. struct drbd_thread *thi = (struct drbd_thread *) arg;
  446. struct drbd_tconn *tconn = thi->tconn;
  447. unsigned long flags;
  448. int retval;
  449. snprintf(current->comm, sizeof(current->comm), "drbd_%c_%s",
  450. thi->name[0], thi->tconn->name);
  451. restart:
  452. retval = thi->function(thi);
  453. spin_lock_irqsave(&thi->t_lock, flags);
  454. /* if the receiver has been "EXITING", the last thing it did
  455. * was set the conn state to "StandAlone",
  456. * if now a re-connect request comes in, conn state goes C_UNCONNECTED,
  457. * and receiver thread will be "started".
  458. * drbd_thread_start needs to set "RESTARTING" in that case.
  459. * t_state check and assignment needs to be within the same spinlock,
  460. * so either thread_start sees EXITING, and can remap to RESTARTING,
  461. * or thread_start see NONE, and can proceed as normal.
  462. */
  463. if (thi->t_state == RESTARTING) {
  464. conn_info(tconn, "Restarting %s thread\n", thi->name);
  465. thi->t_state = RUNNING;
  466. spin_unlock_irqrestore(&thi->t_lock, flags);
  467. goto restart;
  468. }
  469. thi->task = NULL;
  470. thi->t_state = NONE;
  471. smp_mb();
  472. complete_all(&thi->stop);
  473. spin_unlock_irqrestore(&thi->t_lock, flags);
  474. conn_info(tconn, "Terminating %s\n", current->comm);
  475. /* Release mod reference taken when thread was started */
  476. kref_put(&tconn->kref, &conn_destroy);
  477. module_put(THIS_MODULE);
  478. return retval;
  479. }
  480. static void drbd_thread_init(struct drbd_tconn *tconn, struct drbd_thread *thi,
  481. int (*func) (struct drbd_thread *), char *name)
  482. {
  483. spin_lock_init(&thi->t_lock);
  484. thi->task = NULL;
  485. thi->t_state = NONE;
  486. thi->function = func;
  487. thi->tconn = tconn;
  488. strncpy(thi->name, name, ARRAY_SIZE(thi->name));
  489. }
  490. int drbd_thread_start(struct drbd_thread *thi)
  491. {
  492. struct drbd_tconn *tconn = thi->tconn;
  493. struct task_struct *nt;
  494. unsigned long flags;
  495. /* is used from state engine doing drbd_thread_stop_nowait,
  496. * while holding the req lock irqsave */
  497. spin_lock_irqsave(&thi->t_lock, flags);
  498. switch (thi->t_state) {
  499. case NONE:
  500. conn_info(tconn, "Starting %s thread (from %s [%d])\n",
  501. thi->name, current->comm, current->pid);
  502. /* Get ref on module for thread - this is released when thread exits */
  503. if (!try_module_get(THIS_MODULE)) {
  504. conn_err(tconn, "Failed to get module reference in drbd_thread_start\n");
  505. spin_unlock_irqrestore(&thi->t_lock, flags);
  506. return false;
  507. }
  508. kref_get(&thi->tconn->kref);
  509. init_completion(&thi->stop);
  510. thi->reset_cpu_mask = 1;
  511. thi->t_state = RUNNING;
  512. spin_unlock_irqrestore(&thi->t_lock, flags);
  513. flush_signals(current); /* otherw. may get -ERESTARTNOINTR */
  514. nt = kthread_create(drbd_thread_setup, (void *) thi,
  515. "drbd_%c_%s", thi->name[0], thi->tconn->name);
  516. if (IS_ERR(nt)) {
  517. conn_err(tconn, "Couldn't start thread\n");
  518. kref_put(&tconn->kref, &conn_destroy);
  519. module_put(THIS_MODULE);
  520. return false;
  521. }
  522. spin_lock_irqsave(&thi->t_lock, flags);
  523. thi->task = nt;
  524. thi->t_state = RUNNING;
  525. spin_unlock_irqrestore(&thi->t_lock, flags);
  526. wake_up_process(nt);
  527. break;
  528. case EXITING:
  529. thi->t_state = RESTARTING;
  530. conn_info(tconn, "Restarting %s thread (from %s [%d])\n",
  531. thi->name, current->comm, current->pid);
  532. /* fall through */
  533. case RUNNING:
  534. case RESTARTING:
  535. default:
  536. spin_unlock_irqrestore(&thi->t_lock, flags);
  537. break;
  538. }
  539. return true;
  540. }
  541. void _drbd_thread_stop(struct drbd_thread *thi, int restart, int wait)
  542. {
  543. unsigned long flags;
  544. enum drbd_thread_state ns = restart ? RESTARTING : EXITING;
  545. /* may be called from state engine, holding the req lock irqsave */
  546. spin_lock_irqsave(&thi->t_lock, flags);
  547. if (thi->t_state == NONE) {
  548. spin_unlock_irqrestore(&thi->t_lock, flags);
  549. if (restart)
  550. drbd_thread_start(thi);
  551. return;
  552. }
  553. if (thi->t_state != ns) {
  554. if (thi->task == NULL) {
  555. spin_unlock_irqrestore(&thi->t_lock, flags);
  556. return;
  557. }
  558. thi->t_state = ns;
  559. smp_mb();
  560. init_completion(&thi->stop);
  561. if (thi->task != current)
  562. force_sig(DRBD_SIGKILL, thi->task);
  563. }
  564. spin_unlock_irqrestore(&thi->t_lock, flags);
  565. if (wait)
  566. wait_for_completion(&thi->stop);
  567. }
  568. static struct drbd_thread *drbd_task_to_thread(struct drbd_tconn *tconn, struct task_struct *task)
  569. {
  570. struct drbd_thread *thi =
  571. task == tconn->receiver.task ? &tconn->receiver :
  572. task == tconn->asender.task ? &tconn->asender :
  573. task == tconn->worker.task ? &tconn->worker : NULL;
  574. return thi;
  575. }
  576. char *drbd_task_to_thread_name(struct drbd_tconn *tconn, struct task_struct *task)
  577. {
  578. struct drbd_thread *thi = drbd_task_to_thread(tconn, task);
  579. return thi ? thi->name : task->comm;
  580. }
  581. int conn_lowest_minor(struct drbd_tconn *tconn)
  582. {
  583. struct drbd_conf *mdev;
  584. int vnr = 0, m;
  585. rcu_read_lock();
  586. mdev = idr_get_next(&tconn->volumes, &vnr);
  587. m = mdev ? mdev_to_minor(mdev) : -1;
  588. rcu_read_unlock();
  589. return m;
  590. }
  591. #ifdef CONFIG_SMP
  592. /**
  593. * drbd_calc_cpu_mask() - Generate CPU masks, spread over all CPUs
  594. * @mdev: DRBD device.
  595. *
  596. * Forces all threads of a device onto the same CPU. This is beneficial for
  597. * DRBD's performance. May be overwritten by user's configuration.
  598. */
  599. void drbd_calc_cpu_mask(struct drbd_tconn *tconn)
  600. {
  601. int ord, cpu;
  602. /* user override. */
  603. if (cpumask_weight(tconn->cpu_mask))
  604. return;
  605. ord = conn_lowest_minor(tconn) % cpumask_weight(cpu_online_mask);
  606. for_each_online_cpu(cpu) {
  607. if (ord-- == 0) {
  608. cpumask_set_cpu(cpu, tconn->cpu_mask);
  609. return;
  610. }
  611. }
  612. /* should not be reached */
  613. cpumask_setall(tconn->cpu_mask);
  614. }
  615. /**
  616. * drbd_thread_current_set_cpu() - modifies the cpu mask of the _current_ thread
  617. * @mdev: DRBD device.
  618. * @thi: drbd_thread object
  619. *
  620. * call in the "main loop" of _all_ threads, no need for any mutex, current won't die
  621. * prematurely.
  622. */
  623. void drbd_thread_current_set_cpu(struct drbd_thread *thi)
  624. {
  625. struct task_struct *p = current;
  626. if (!thi->reset_cpu_mask)
  627. return;
  628. thi->reset_cpu_mask = 0;
  629. set_cpus_allowed_ptr(p, thi->tconn->cpu_mask);
  630. }
  631. #endif
  632. /**
  633. * drbd_header_size - size of a packet header
  634. *
  635. * The header size is a multiple of 8, so any payload following the header is
  636. * word aligned on 64-bit architectures. (The bitmap send and receive code
  637. * relies on this.)
  638. */
  639. unsigned int drbd_header_size(struct drbd_tconn *tconn)
  640. {
  641. if (tconn->agreed_pro_version >= 100) {
  642. BUILD_BUG_ON(!IS_ALIGNED(sizeof(struct p_header100), 8));
  643. return sizeof(struct p_header100);
  644. } else {
  645. BUILD_BUG_ON(sizeof(struct p_header80) !=
  646. sizeof(struct p_header95));
  647. BUILD_BUG_ON(!IS_ALIGNED(sizeof(struct p_header80), 8));
  648. return sizeof(struct p_header80);
  649. }
  650. }
  651. static unsigned int prepare_header80(struct p_header80 *h, enum drbd_packet cmd, int size)
  652. {
  653. h->magic = cpu_to_be32(DRBD_MAGIC);
  654. h->command = cpu_to_be16(cmd);
  655. h->length = cpu_to_be16(size);
  656. return sizeof(struct p_header80);
  657. }
  658. static unsigned int prepare_header95(struct p_header95 *h, enum drbd_packet cmd, int size)
  659. {
  660. h->magic = cpu_to_be16(DRBD_MAGIC_BIG);
  661. h->command = cpu_to_be16(cmd);
  662. h->length = cpu_to_be32(size);
  663. return sizeof(struct p_header95);
  664. }
  665. static unsigned int prepare_header100(struct p_header100 *h, enum drbd_packet cmd,
  666. int size, int vnr)
  667. {
  668. h->magic = cpu_to_be32(DRBD_MAGIC_100);
  669. h->volume = cpu_to_be16(vnr);
  670. h->command = cpu_to_be16(cmd);
  671. h->length = cpu_to_be32(size);
  672. h->pad = 0;
  673. return sizeof(struct p_header100);
  674. }
  675. static unsigned int prepare_header(struct drbd_tconn *tconn, int vnr,
  676. void *buffer, enum drbd_packet cmd, int size)
  677. {
  678. if (tconn->agreed_pro_version >= 100)
  679. return prepare_header100(buffer, cmd, size, vnr);
  680. else if (tconn->agreed_pro_version >= 95 &&
  681. size > DRBD_MAX_SIZE_H80_PACKET)
  682. return prepare_header95(buffer, cmd, size);
  683. else
  684. return prepare_header80(buffer, cmd, size);
  685. }
  686. static void *__conn_prepare_command(struct drbd_tconn *tconn,
  687. struct drbd_socket *sock)
  688. {
  689. if (!sock->socket)
  690. return NULL;
  691. return sock->sbuf + drbd_header_size(tconn);
  692. }
  693. void *conn_prepare_command(struct drbd_tconn *tconn, struct drbd_socket *sock)
  694. {
  695. void *p;
  696. mutex_lock(&sock->mutex);
  697. p = __conn_prepare_command(tconn, sock);
  698. if (!p)
  699. mutex_unlock(&sock->mutex);
  700. return p;
  701. }
  702. void *drbd_prepare_command(struct drbd_conf *mdev, struct drbd_socket *sock)
  703. {
  704. return conn_prepare_command(mdev->tconn, sock);
  705. }
  706. static int __send_command(struct drbd_tconn *tconn, int vnr,
  707. struct drbd_socket *sock, enum drbd_packet cmd,
  708. unsigned int header_size, void *data,
  709. unsigned int size)
  710. {
  711. int msg_flags;
  712. int err;
  713. /*
  714. * Called with @data == NULL and the size of the data blocks in @size
  715. * for commands that send data blocks. For those commands, omit the
  716. * MSG_MORE flag: this will increase the likelihood that data blocks
  717. * which are page aligned on the sender will end up page aligned on the
  718. * receiver.
  719. */
  720. msg_flags = data ? MSG_MORE : 0;
  721. header_size += prepare_header(tconn, vnr, sock->sbuf, cmd,
  722. header_size + size);
  723. err = drbd_send_all(tconn, sock->socket, sock->sbuf, header_size,
  724. msg_flags);
  725. if (data && !err)
  726. err = drbd_send_all(tconn, sock->socket, data, size, 0);
  727. return err;
  728. }
  729. static int __conn_send_command(struct drbd_tconn *tconn, struct drbd_socket *sock,
  730. enum drbd_packet cmd, unsigned int header_size,
  731. void *data, unsigned int size)
  732. {
  733. return __send_command(tconn, 0, sock, cmd, header_size, data, size);
  734. }
  735. int conn_send_command(struct drbd_tconn *tconn, struct drbd_socket *sock,
  736. enum drbd_packet cmd, unsigned int header_size,
  737. void *data, unsigned int size)
  738. {
  739. int err;
  740. err = __conn_send_command(tconn, sock, cmd, header_size, data, size);
  741. mutex_unlock(&sock->mutex);
  742. return err;
  743. }
  744. int drbd_send_command(struct drbd_conf *mdev, struct drbd_socket *sock,
  745. enum drbd_packet cmd, unsigned int header_size,
  746. void *data, unsigned int size)
  747. {
  748. int err;
  749. err = __send_command(mdev->tconn, mdev->vnr, sock, cmd, header_size,
  750. data, size);
  751. mutex_unlock(&sock->mutex);
  752. return err;
  753. }
  754. int drbd_send_ping(struct drbd_tconn *tconn)
  755. {
  756. struct drbd_socket *sock;
  757. sock = &tconn->meta;
  758. if (!conn_prepare_command(tconn, sock))
  759. return -EIO;
  760. return conn_send_command(tconn, sock, P_PING, 0, NULL, 0);
  761. }
  762. int drbd_send_ping_ack(struct drbd_tconn *tconn)
  763. {
  764. struct drbd_socket *sock;
  765. sock = &tconn->meta;
  766. if (!conn_prepare_command(tconn, sock))
  767. return -EIO;
  768. return conn_send_command(tconn, sock, P_PING_ACK, 0, NULL, 0);
  769. }
  770. int drbd_send_sync_param(struct drbd_conf *mdev)
  771. {
  772. struct drbd_socket *sock;
  773. struct p_rs_param_95 *p;
  774. int size;
  775. const int apv = mdev->tconn->agreed_pro_version;
  776. enum drbd_packet cmd;
  777. struct net_conf *nc;
  778. struct disk_conf *dc;
  779. sock = &mdev->tconn->data;
  780. p = drbd_prepare_command(mdev, sock);
  781. if (!p)
  782. return -EIO;
  783. rcu_read_lock();
  784. nc = rcu_dereference(mdev->tconn->net_conf);
  785. size = apv <= 87 ? sizeof(struct p_rs_param)
  786. : apv == 88 ? sizeof(struct p_rs_param)
  787. + strlen(nc->verify_alg) + 1
  788. : apv <= 94 ? sizeof(struct p_rs_param_89)
  789. : /* apv >= 95 */ sizeof(struct p_rs_param_95);
  790. cmd = apv >= 89 ? P_SYNC_PARAM89 : P_SYNC_PARAM;
  791. /* initialize verify_alg and csums_alg */
  792. memset(p->verify_alg, 0, 2 * SHARED_SECRET_MAX);
  793. if (get_ldev(mdev)) {
  794. dc = rcu_dereference(mdev->ldev->disk_conf);
  795. p->resync_rate = cpu_to_be32(dc->resync_rate);
  796. p->c_plan_ahead = cpu_to_be32(dc->c_plan_ahead);
  797. p->c_delay_target = cpu_to_be32(dc->c_delay_target);
  798. p->c_fill_target = cpu_to_be32(dc->c_fill_target);
  799. p->c_max_rate = cpu_to_be32(dc->c_max_rate);
  800. put_ldev(mdev);
  801. } else {
  802. p->resync_rate = cpu_to_be32(DRBD_RESYNC_RATE_DEF);
  803. p->c_plan_ahead = cpu_to_be32(DRBD_C_PLAN_AHEAD_DEF);
  804. p->c_delay_target = cpu_to_be32(DRBD_C_DELAY_TARGET_DEF);
  805. p->c_fill_target = cpu_to_be32(DRBD_C_FILL_TARGET_DEF);
  806. p->c_max_rate = cpu_to_be32(DRBD_C_MAX_RATE_DEF);
  807. }
  808. if (apv >= 88)
  809. strcpy(p->verify_alg, nc->verify_alg);
  810. if (apv >= 89)
  811. strcpy(p->csums_alg, nc->csums_alg);
  812. rcu_read_unlock();
  813. return drbd_send_command(mdev, sock, cmd, size, NULL, 0);
  814. }
  815. int __drbd_send_protocol(struct drbd_tconn *tconn, enum drbd_packet cmd)
  816. {
  817. struct drbd_socket *sock;
  818. struct p_protocol *p;
  819. struct net_conf *nc;
  820. int size, cf;
  821. sock = &tconn->data;
  822. p = __conn_prepare_command(tconn, sock);
  823. if (!p)
  824. return -EIO;
  825. rcu_read_lock();
  826. nc = rcu_dereference(tconn->net_conf);
  827. if (nc->tentative && tconn->agreed_pro_version < 92) {
  828. rcu_read_unlock();
  829. mutex_unlock(&sock->mutex);
  830. conn_err(tconn, "--dry-run is not supported by peer");
  831. return -EOPNOTSUPP;
  832. }
  833. size = sizeof(*p);
  834. if (tconn->agreed_pro_version >= 87)
  835. size += strlen(nc->integrity_alg) + 1;
  836. p->protocol = cpu_to_be32(nc->wire_protocol);
  837. p->after_sb_0p = cpu_to_be32(nc->after_sb_0p);
  838. p->after_sb_1p = cpu_to_be32(nc->after_sb_1p);
  839. p->after_sb_2p = cpu_to_be32(nc->after_sb_2p);
  840. p->two_primaries = cpu_to_be32(nc->two_primaries);
  841. cf = 0;
  842. if (nc->discard_my_data)
  843. cf |= CF_DISCARD_MY_DATA;
  844. if (nc->tentative)
  845. cf |= CF_DRY_RUN;
  846. p->conn_flags = cpu_to_be32(cf);
  847. if (tconn->agreed_pro_version >= 87)
  848. strcpy(p->integrity_alg, nc->integrity_alg);
  849. rcu_read_unlock();
  850. return __conn_send_command(tconn, sock, cmd, size, NULL, 0);
  851. }
  852. int drbd_send_protocol(struct drbd_tconn *tconn)
  853. {
  854. int err;
  855. mutex_lock(&tconn->data.mutex);
  856. err = __drbd_send_protocol(tconn, P_PROTOCOL);
  857. mutex_unlock(&tconn->data.mutex);
  858. return err;
  859. }
  860. int _drbd_send_uuids(struct drbd_conf *mdev, u64 uuid_flags)
  861. {
  862. struct drbd_socket *sock;
  863. struct p_uuids *p;
  864. int i;
  865. if (!get_ldev_if_state(mdev, D_NEGOTIATING))
  866. return 0;
  867. sock = &mdev->tconn->data;
  868. p = drbd_prepare_command(mdev, sock);
  869. if (!p) {
  870. put_ldev(mdev);
  871. return -EIO;
  872. }
  873. for (i = UI_CURRENT; i < UI_SIZE; i++)
  874. p->uuid[i] = mdev->ldev ? cpu_to_be64(mdev->ldev->md.uuid[i]) : 0;
  875. mdev->comm_bm_set = drbd_bm_total_weight(mdev);
  876. p->uuid[UI_SIZE] = cpu_to_be64(mdev->comm_bm_set);
  877. rcu_read_lock();
  878. uuid_flags |= rcu_dereference(mdev->tconn->net_conf)->discard_my_data ? 1 : 0;
  879. rcu_read_unlock();
  880. uuid_flags |= test_bit(CRASHED_PRIMARY, &mdev->flags) ? 2 : 0;
  881. uuid_flags |= mdev->new_state_tmp.disk == D_INCONSISTENT ? 4 : 0;
  882. p->uuid[UI_FLAGS] = cpu_to_be64(uuid_flags);
  883. put_ldev(mdev);
  884. return drbd_send_command(mdev, sock, P_UUIDS, sizeof(*p), NULL, 0);
  885. }
  886. int drbd_send_uuids(struct drbd_conf *mdev)
  887. {
  888. return _drbd_send_uuids(mdev, 0);
  889. }
  890. int drbd_send_uuids_skip_initial_sync(struct drbd_conf *mdev)
  891. {
  892. return _drbd_send_uuids(mdev, 8);
  893. }
  894. void drbd_print_uuids(struct drbd_conf *mdev, const char *text)
  895. {
  896. if (get_ldev_if_state(mdev, D_NEGOTIATING)) {
  897. u64 *uuid = mdev->ldev->md.uuid;
  898. dev_info(DEV, "%s %016llX:%016llX:%016llX:%016llX\n",
  899. text,
  900. (unsigned long long)uuid[UI_CURRENT],
  901. (unsigned long long)uuid[UI_BITMAP],
  902. (unsigned long long)uuid[UI_HISTORY_START],
  903. (unsigned long long)uuid[UI_HISTORY_END]);
  904. put_ldev(mdev);
  905. } else {
  906. dev_info(DEV, "%s effective data uuid: %016llX\n",
  907. text,
  908. (unsigned long long)mdev->ed_uuid);
  909. }
  910. }
  911. void drbd_gen_and_send_sync_uuid(struct drbd_conf *mdev)
  912. {
  913. struct drbd_socket *sock;
  914. struct p_rs_uuid *p;
  915. u64 uuid;
  916. D_ASSERT(mdev->state.disk == D_UP_TO_DATE);
  917. uuid = mdev->ldev->md.uuid[UI_BITMAP];
  918. if (uuid && uuid != UUID_JUST_CREATED)
  919. uuid = uuid + UUID_NEW_BM_OFFSET;
  920. else
  921. get_random_bytes(&uuid, sizeof(u64));
  922. drbd_uuid_set(mdev, UI_BITMAP, uuid);
  923. drbd_print_uuids(mdev, "updated sync UUID");
  924. drbd_md_sync(mdev);
  925. sock = &mdev->tconn->data;
  926. p = drbd_prepare_command(mdev, sock);
  927. if (p) {
  928. p->uuid = cpu_to_be64(uuid);
  929. drbd_send_command(mdev, sock, P_SYNC_UUID, sizeof(*p), NULL, 0);
  930. }
  931. }
  932. int drbd_send_sizes(struct drbd_conf *mdev, int trigger_reply, enum dds_flags flags)
  933. {
  934. struct drbd_socket *sock;
  935. struct p_sizes *p;
  936. sector_t d_size, u_size;
  937. int q_order_type, max_bio_size;
  938. if (get_ldev_if_state(mdev, D_NEGOTIATING)) {
  939. D_ASSERT(mdev->ldev->backing_bdev);
  940. d_size = drbd_get_max_capacity(mdev->ldev);
  941. rcu_read_lock();
  942. u_size = rcu_dereference(mdev->ldev->disk_conf)->disk_size;
  943. rcu_read_unlock();
  944. q_order_type = drbd_queue_order_type(mdev);
  945. max_bio_size = queue_max_hw_sectors(mdev->ldev->backing_bdev->bd_disk->queue) << 9;
  946. max_bio_size = min_t(int, max_bio_size, DRBD_MAX_BIO_SIZE);
  947. put_ldev(mdev);
  948. } else {
  949. d_size = 0;
  950. u_size = 0;
  951. q_order_type = QUEUE_ORDERED_NONE;
  952. max_bio_size = DRBD_MAX_BIO_SIZE; /* ... multiple BIOs per peer_request */
  953. }
  954. sock = &mdev->tconn->data;
  955. p = drbd_prepare_command(mdev, sock);
  956. if (!p)
  957. return -EIO;
  958. if (mdev->tconn->agreed_pro_version <= 94)
  959. max_bio_size = min_t(int, max_bio_size, DRBD_MAX_SIZE_H80_PACKET);
  960. else if (mdev->tconn->agreed_pro_version < 100)
  961. max_bio_size = min_t(int, max_bio_size, DRBD_MAX_BIO_SIZE_P95);
  962. p->d_size = cpu_to_be64(d_size);
  963. p->u_size = cpu_to_be64(u_size);
  964. p->c_size = cpu_to_be64(trigger_reply ? 0 : drbd_get_capacity(mdev->this_bdev));
  965. p->max_bio_size = cpu_to_be32(max_bio_size);
  966. p->queue_order_type = cpu_to_be16(q_order_type);
  967. p->dds_flags = cpu_to_be16(flags);
  968. return drbd_send_command(mdev, sock, P_SIZES, sizeof(*p), NULL, 0);
  969. }
  970. /**
  971. * drbd_send_current_state() - Sends the drbd state to the peer
  972. * @mdev: DRBD device.
  973. */
  974. int drbd_send_current_state(struct drbd_conf *mdev)
  975. {
  976. struct drbd_socket *sock;
  977. struct p_state *p;
  978. sock = &mdev->tconn->data;
  979. p = drbd_prepare_command(mdev, sock);
  980. if (!p)
  981. return -EIO;
  982. p->state = cpu_to_be32(mdev->state.i); /* Within the send mutex */
  983. return drbd_send_command(mdev, sock, P_STATE, sizeof(*p), NULL, 0);
  984. }
  985. /**
  986. * drbd_send_state() - After a state change, sends the new state to the peer
  987. * @mdev: DRBD device.
  988. * @state: the state to send, not necessarily the current state.
  989. *
  990. * Each state change queues an "after_state_ch" work, which will eventually
  991. * send the resulting new state to the peer. If more state changes happen
  992. * between queuing and processing of the after_state_ch work, we still
  993. * want to send each intermediary state in the order it occurred.
  994. */
  995. int drbd_send_state(struct drbd_conf *mdev, union drbd_state state)
  996. {
  997. struct drbd_socket *sock;
  998. struct p_state *p;
  999. sock = &mdev->tconn->data;
  1000. p = drbd_prepare_command(mdev, sock);
  1001. if (!p)
  1002. return -EIO;
  1003. p->state = cpu_to_be32(state.i); /* Within the send mutex */
  1004. return drbd_send_command(mdev, sock, P_STATE, sizeof(*p), NULL, 0);
  1005. }
  1006. int drbd_send_state_req(struct drbd_conf *mdev, union drbd_state mask, union drbd_state val)
  1007. {
  1008. struct drbd_socket *sock;
  1009. struct p_req_state *p;
  1010. sock = &mdev->tconn->data;
  1011. p = drbd_prepare_command(mdev, sock);
  1012. if (!p)
  1013. return -EIO;
  1014. p->mask = cpu_to_be32(mask.i);
  1015. p->val = cpu_to_be32(val.i);
  1016. return drbd_send_command(mdev, sock, P_STATE_CHG_REQ, sizeof(*p), NULL, 0);
  1017. }
  1018. int conn_send_state_req(struct drbd_tconn *tconn, union drbd_state mask, union drbd_state val)
  1019. {
  1020. enum drbd_packet cmd;
  1021. struct drbd_socket *sock;
  1022. struct p_req_state *p;
  1023. cmd = tconn->agreed_pro_version < 100 ? P_STATE_CHG_REQ : P_CONN_ST_CHG_REQ;
  1024. sock = &tconn->data;
  1025. p = conn_prepare_command(tconn, sock);
  1026. if (!p)
  1027. return -EIO;
  1028. p->mask = cpu_to_be32(mask.i);
  1029. p->val = cpu_to_be32(val.i);
  1030. return conn_send_command(tconn, sock, cmd, sizeof(*p), NULL, 0);
  1031. }
  1032. void drbd_send_sr_reply(struct drbd_conf *mdev, enum drbd_state_rv retcode)
  1033. {
  1034. struct drbd_socket *sock;
  1035. struct p_req_state_reply *p;
  1036. sock = &mdev->tconn->meta;
  1037. p = drbd_prepare_command(mdev, sock);
  1038. if (p) {
  1039. p->retcode = cpu_to_be32(retcode);
  1040. drbd_send_command(mdev, sock, P_STATE_CHG_REPLY, sizeof(*p), NULL, 0);
  1041. }
  1042. }
  1043. void conn_send_sr_reply(struct drbd_tconn *tconn, enum drbd_state_rv retcode)
  1044. {
  1045. struct drbd_socket *sock;
  1046. struct p_req_state_reply *p;
  1047. enum drbd_packet cmd = tconn->agreed_pro_version < 100 ? P_STATE_CHG_REPLY : P_CONN_ST_CHG_REPLY;
  1048. sock = &tconn->meta;
  1049. p = conn_prepare_command(tconn, sock);
  1050. if (p) {
  1051. p->retcode = cpu_to_be32(retcode);
  1052. conn_send_command(tconn, sock, cmd, sizeof(*p), NULL, 0);
  1053. }
  1054. }
  1055. static void dcbp_set_code(struct p_compressed_bm *p, enum drbd_bitmap_code code)
  1056. {
  1057. BUG_ON(code & ~0xf);
  1058. p->encoding = (p->encoding & ~0xf) | code;
  1059. }
  1060. static void dcbp_set_start(struct p_compressed_bm *p, int set)
  1061. {
  1062. p->encoding = (p->encoding & ~0x80) | (set ? 0x80 : 0);
  1063. }
  1064. static void dcbp_set_pad_bits(struct p_compressed_bm *p, int n)
  1065. {
  1066. BUG_ON(n & ~0x7);
  1067. p->encoding = (p->encoding & (~0x7 << 4)) | (n << 4);
  1068. }
  1069. int fill_bitmap_rle_bits(struct drbd_conf *mdev,
  1070. struct p_compressed_bm *p,
  1071. unsigned int size,
  1072. struct bm_xfer_ctx *c)
  1073. {
  1074. struct bitstream bs;
  1075. unsigned long plain_bits;
  1076. unsigned long tmp;
  1077. unsigned long rl;
  1078. unsigned len;
  1079. unsigned toggle;
  1080. int bits, use_rle;
  1081. /* may we use this feature? */
  1082. rcu_read_lock();
  1083. use_rle = rcu_dereference(mdev->tconn->net_conf)->use_rle;
  1084. rcu_read_unlock();
  1085. if (!use_rle || mdev->tconn->agreed_pro_version < 90)
  1086. return 0;
  1087. if (c->bit_offset >= c->bm_bits)
  1088. return 0; /* nothing to do. */
  1089. /* use at most thus many bytes */
  1090. bitstream_init(&bs, p->code, size, 0);
  1091. memset(p->code, 0, size);
  1092. /* plain bits covered in this code string */
  1093. plain_bits = 0;
  1094. /* p->encoding & 0x80 stores whether the first run length is set.
  1095. * bit offset is implicit.
  1096. * start with toggle == 2 to be able to tell the first iteration */
  1097. toggle = 2;
  1098. /* see how much plain bits we can stuff into one packet
  1099. * using RLE and VLI. */
  1100. do {
  1101. tmp = (toggle == 0) ? _drbd_bm_find_next_zero(mdev, c->bit_offset)
  1102. : _drbd_bm_find_next(mdev, c->bit_offset);
  1103. if (tmp == -1UL)
  1104. tmp = c->bm_bits;
  1105. rl = tmp - c->bit_offset;
  1106. if (toggle == 2) { /* first iteration */
  1107. if (rl == 0) {
  1108. /* the first checked bit was set,
  1109. * store start value, */
  1110. dcbp_set_start(p, 1);
  1111. /* but skip encoding of zero run length */
  1112. toggle = !toggle;
  1113. continue;
  1114. }
  1115. dcbp_set_start(p, 0);
  1116. }
  1117. /* paranoia: catch zero runlength.
  1118. * can only happen if bitmap is modified while we scan it. */
  1119. if (rl == 0) {
  1120. dev_err(DEV, "unexpected zero runlength while encoding bitmap "
  1121. "t:%u bo:%lu\n", toggle, c->bit_offset);
  1122. return -1;
  1123. }
  1124. bits = vli_encode_bits(&bs, rl);
  1125. if (bits == -ENOBUFS) /* buffer full */
  1126. break;
  1127. if (bits <= 0) {
  1128. dev_err(DEV, "error while encoding bitmap: %d\n", bits);
  1129. return 0;
  1130. }
  1131. toggle = !toggle;
  1132. plain_bits += rl;
  1133. c->bit_offset = tmp;
  1134. } while (c->bit_offset < c->bm_bits);
  1135. len = bs.cur.b - p->code + !!bs.cur.bit;
  1136. if (plain_bits < (len << 3)) {
  1137. /* incompressible with this method.
  1138. * we need to rewind both word and bit position. */
  1139. c->bit_offset -= plain_bits;
  1140. bm_xfer_ctx_bit_to_word_offset(c);
  1141. c->bit_offset = c->word_offset * BITS_PER_LONG;
  1142. return 0;
  1143. }
  1144. /* RLE + VLI was able to compress it just fine.
  1145. * update c->word_offset. */
  1146. bm_xfer_ctx_bit_to_word_offset(c);
  1147. /* store pad_bits */
  1148. dcbp_set_pad_bits(p, (8 - bs.cur.bit) & 0x7);
  1149. return len;
  1150. }
  1151. /**
  1152. * send_bitmap_rle_or_plain
  1153. *
  1154. * Return 0 when done, 1 when another iteration is needed, and a negative error
  1155. * code upon failure.
  1156. */
  1157. static int
  1158. send_bitmap_rle_or_plain(struct drbd_conf *mdev, struct bm_xfer_ctx *c)
  1159. {
  1160. struct drbd_socket *sock = &mdev->tconn->data;
  1161. unsigned int header_size = drbd_header_size(mdev->tconn);
  1162. struct p_compressed_bm *p = sock->sbuf + header_size;
  1163. int len, err;
  1164. len = fill_bitmap_rle_bits(mdev, p,
  1165. DRBD_SOCKET_BUFFER_SIZE - header_size - sizeof(*p), c);
  1166. if (len < 0)
  1167. return -EIO;
  1168. if (len) {
  1169. dcbp_set_code(p, RLE_VLI_Bits);
  1170. err = __send_command(mdev->tconn, mdev->vnr, sock,
  1171. P_COMPRESSED_BITMAP, sizeof(*p) + len,
  1172. NULL, 0);
  1173. c->packets[0]++;
  1174. c->bytes[0] += header_size + sizeof(*p) + len;
  1175. if (c->bit_offset >= c->bm_bits)
  1176. len = 0; /* DONE */
  1177. } else {
  1178. /* was not compressible.
  1179. * send a buffer full of plain text bits instead. */
  1180. unsigned int data_size;
  1181. unsigned long num_words;
  1182. unsigned long *p = sock->sbuf + header_size;
  1183. data_size = DRBD_SOCKET_BUFFER_SIZE - header_size;
  1184. num_words = min_t(size_t, data_size / sizeof(*p),
  1185. c->bm_words - c->word_offset);
  1186. len = num_words * sizeof(*p);
  1187. if (len)
  1188. drbd_bm_get_lel(mdev, c->word_offset, num_words, p);
  1189. err = __send_command(mdev->tconn, mdev->vnr, sock, P_BITMAP, len, NULL, 0);
  1190. c->word_offset += num_words;
  1191. c->bit_offset = c->word_offset * BITS_PER_LONG;
  1192. c->packets[1]++;
  1193. c->bytes[1] += header_size + len;
  1194. if (c->bit_offset > c->bm_bits)
  1195. c->bit_offset = c->bm_bits;
  1196. }
  1197. if (!err) {
  1198. if (len == 0) {
  1199. INFO_bm_xfer_stats(mdev, "send", c);
  1200. return 0;
  1201. } else
  1202. return 1;
  1203. }
  1204. return -EIO;
  1205. }
  1206. /* See the comment at receive_bitmap() */
  1207. static int _drbd_send_bitmap(struct drbd_conf *mdev)
  1208. {
  1209. struct bm_xfer_ctx c;
  1210. int err;
  1211. if (!expect(mdev->bitmap))
  1212. return false;
  1213. if (get_ldev(mdev)) {
  1214. if (drbd_md_test_flag(mdev->ldev, MDF_FULL_SYNC)) {
  1215. dev_info(DEV, "Writing the whole bitmap, MDF_FullSync was set.\n");
  1216. drbd_bm_set_all(mdev);
  1217. if (drbd_bm_write(mdev)) {
  1218. /* write_bm did fail! Leave full sync flag set in Meta P_DATA
  1219. * but otherwise process as per normal - need to tell other
  1220. * side that a full resync is required! */
  1221. dev_err(DEV, "Failed to write bitmap to disk!\n");
  1222. } else {
  1223. drbd_md_clear_flag(mdev, MDF_FULL_SYNC);
  1224. drbd_md_sync(mdev);
  1225. }
  1226. }
  1227. put_ldev(mdev);
  1228. }
  1229. c = (struct bm_xfer_ctx) {
  1230. .bm_bits = drbd_bm_bits(mdev),
  1231. .bm_words = drbd_bm_words(mdev),
  1232. };
  1233. do {
  1234. err = send_bitmap_rle_or_plain(mdev, &c);
  1235. } while (err > 0);
  1236. return err == 0;
  1237. }
  1238. int drbd_send_bitmap(struct drbd_conf *mdev)
  1239. {
  1240. struct drbd_socket *sock = &mdev->tconn->data;
  1241. int err = -1;
  1242. mutex_lock(&sock->mutex);
  1243. if (sock->socket)
  1244. err = !_drbd_send_bitmap(mdev);
  1245. mutex_unlock(&sock->mutex);
  1246. return err;
  1247. }
  1248. void drbd_send_b_ack(struct drbd_tconn *tconn, u32 barrier_nr, u32 set_size)
  1249. {
  1250. struct drbd_socket *sock;
  1251. struct p_barrier_ack *p;
  1252. if (tconn->cstate < C_WF_REPORT_PARAMS)
  1253. return;
  1254. sock = &tconn->meta;
  1255. p = conn_prepare_command(tconn, sock);
  1256. if (!p)
  1257. return;
  1258. p->barrier = barrier_nr;
  1259. p->set_size = cpu_to_be32(set_size);
  1260. conn_send_command(tconn, sock, P_BARRIER_ACK, sizeof(*p), NULL, 0);
  1261. }
  1262. /**
  1263. * _drbd_send_ack() - Sends an ack packet
  1264. * @mdev: DRBD device.
  1265. * @cmd: Packet command code.
  1266. * @sector: sector, needs to be in big endian byte order
  1267. * @blksize: size in byte, needs to be in big endian byte order
  1268. * @block_id: Id, big endian byte order
  1269. */
  1270. static int _drbd_send_ack(struct drbd_conf *mdev, enum drbd_packet cmd,
  1271. u64 sector, u32 blksize, u64 block_id)
  1272. {
  1273. struct drbd_socket *sock;
  1274. struct p_block_ack *p;
  1275. if (mdev->state.conn < C_CONNECTED)
  1276. return -EIO;
  1277. sock = &mdev->tconn->meta;
  1278. p = drbd_prepare_command(mdev, sock);
  1279. if (!p)
  1280. return -EIO;
  1281. p->sector = sector;
  1282. p->block_id = block_id;
  1283. p->blksize = blksize;
  1284. p->seq_num = cpu_to_be32(atomic_inc_return(&mdev->packet_seq));
  1285. return drbd_send_command(mdev, sock, cmd, sizeof(*p), NULL, 0);
  1286. }
  1287. /* dp->sector and dp->block_id already/still in network byte order,
  1288. * data_size is payload size according to dp->head,
  1289. * and may need to be corrected for digest size. */
  1290. void drbd_send_ack_dp(struct drbd_conf *mdev, enum drbd_packet cmd,
  1291. struct p_data *dp, int data_size)
  1292. {
  1293. if (mdev->tconn->peer_integrity_tfm)
  1294. data_size -= crypto_hash_digestsize(mdev->tconn->peer_integrity_tfm);
  1295. _drbd_send_ack(mdev, cmd, dp->sector, cpu_to_be32(data_size),
  1296. dp->block_id);
  1297. }
  1298. void drbd_send_ack_rp(struct drbd_conf *mdev, enum drbd_packet cmd,
  1299. struct p_block_req *rp)
  1300. {
  1301. _drbd_send_ack(mdev, cmd, rp->sector, rp->blksize, rp->block_id);
  1302. }
  1303. /**
  1304. * drbd_send_ack() - Sends an ack packet
  1305. * @mdev: DRBD device
  1306. * @cmd: packet command code
  1307. * @peer_req: peer request
  1308. */
  1309. int drbd_send_ack(struct drbd_conf *mdev, enum drbd_packet cmd,
  1310. struct drbd_peer_request *peer_req)
  1311. {
  1312. return _drbd_send_ack(mdev, cmd,
  1313. cpu_to_be64(peer_req->i.sector),
  1314. cpu_to_be32(peer_req->i.size),
  1315. peer_req->block_id);
  1316. }
  1317. /* This function misuses the block_id field to signal if the blocks
  1318. * are is sync or not. */
  1319. int drbd_send_ack_ex(struct drbd_conf *mdev, enum drbd_packet cmd,
  1320. sector_t sector, int blksize, u64 block_id)
  1321. {
  1322. return _drbd_send_ack(mdev, cmd,
  1323. cpu_to_be64(sector),
  1324. cpu_to_be32(blksize),
  1325. cpu_to_be64(block_id));
  1326. }
  1327. int drbd_send_drequest(struct drbd_conf *mdev, int cmd,
  1328. sector_t sector, int size, u64 block_id)
  1329. {
  1330. struct drbd_socket *sock;
  1331. struct p_block_req *p;
  1332. sock = &mdev->tconn->data;
  1333. p = drbd_prepare_command(mdev, sock);
  1334. if (!p)
  1335. return -EIO;
  1336. p->sector = cpu_to_be64(sector);
  1337. p->block_id = block_id;
  1338. p->blksize = cpu_to_be32(size);
  1339. return drbd_send_command(mdev, sock, cmd, sizeof(*p), NULL, 0);
  1340. }
  1341. int drbd_send_drequest_csum(struct drbd_conf *mdev, sector_t sector, int size,
  1342. void *digest, int digest_size, enum drbd_packet cmd)
  1343. {
  1344. struct drbd_socket *sock;
  1345. struct p_block_req *p;
  1346. /* FIXME: Put the digest into the preallocated socket buffer. */
  1347. sock = &mdev->tconn->data;
  1348. p = drbd_prepare_command(mdev, sock);
  1349. if (!p)
  1350. return -EIO;
  1351. p->sector = cpu_to_be64(sector);
  1352. p->block_id = ID_SYNCER /* unused */;
  1353. p->blksize = cpu_to_be32(size);
  1354. return drbd_send_command(mdev, sock, cmd, sizeof(*p),
  1355. digest, digest_size);
  1356. }
  1357. int drbd_send_ov_request(struct drbd_conf *mdev, sector_t sector, int size)
  1358. {
  1359. struct drbd_socket *sock;
  1360. struct p_block_req *p;
  1361. sock = &mdev->tconn->data;
  1362. p = drbd_prepare_command(mdev, sock);
  1363. if (!p)
  1364. return -EIO;
  1365. p->sector = cpu_to_be64(sector);
  1366. p->block_id = ID_SYNCER /* unused */;
  1367. p->blksize = cpu_to_be32(size);
  1368. return drbd_send_command(mdev, sock, P_OV_REQUEST, sizeof(*p), NULL, 0);
  1369. }
  1370. /* called on sndtimeo
  1371. * returns false if we should retry,
  1372. * true if we think connection is dead
  1373. */
  1374. static int we_should_drop_the_connection(struct drbd_tconn *tconn, struct socket *sock)
  1375. {
  1376. int drop_it;
  1377. /* long elapsed = (long)(jiffies - mdev->last_received); */
  1378. drop_it = tconn->meta.socket == sock
  1379. || !tconn->asender.task
  1380. || get_t_state(&tconn->asender) != RUNNING
  1381. || tconn->cstate < C_WF_REPORT_PARAMS;
  1382. if (drop_it)
  1383. return true;
  1384. drop_it = !--tconn->ko_count;
  1385. if (!drop_it) {
  1386. conn_err(tconn, "[%s/%d] sock_sendmsg time expired, ko = %u\n",
  1387. current->comm, current->pid, tconn->ko_count);
  1388. request_ping(tconn);
  1389. }
  1390. return drop_it; /* && (mdev->state == R_PRIMARY) */;
  1391. }
  1392. static void drbd_update_congested(struct drbd_tconn *tconn)
  1393. {
  1394. struct sock *sk = tconn->data.socket->sk;
  1395. if (sk->sk_wmem_queued > sk->sk_sndbuf * 4 / 5)
  1396. set_bit(NET_CONGESTED, &tconn->flags);
  1397. }
  1398. /* The idea of sendpage seems to be to put some kind of reference
  1399. * to the page into the skb, and to hand it over to the NIC. In
  1400. * this process get_page() gets called.
  1401. *
  1402. * As soon as the page was really sent over the network put_page()
  1403. * gets called by some part of the network layer. [ NIC driver? ]
  1404. *
  1405. * [ get_page() / put_page() increment/decrement the count. If count
  1406. * reaches 0 the page will be freed. ]
  1407. *
  1408. * This works nicely with pages from FSs.
  1409. * But this means that in protocol A we might signal IO completion too early!
  1410. *
  1411. * In order not to corrupt data during a resync we must make sure
  1412. * that we do not reuse our own buffer pages (EEs) to early, therefore
  1413. * we have the net_ee list.
  1414. *
  1415. * XFS seems to have problems, still, it submits pages with page_count == 0!
  1416. * As a workaround, we disable sendpage on pages
  1417. * with page_count == 0 or PageSlab.
  1418. */
  1419. static int _drbd_no_send_page(struct drbd_conf *mdev, struct page *page,
  1420. int offset, size_t size, unsigned msg_flags)
  1421. {
  1422. struct socket *socket;
  1423. void *addr;
  1424. int err;
  1425. socket = mdev->tconn->data.socket;
  1426. addr = kmap(page) + offset;
  1427. err = drbd_send_all(mdev->tconn, socket, addr, size, msg_flags);
  1428. kunmap(page);
  1429. if (!err)
  1430. mdev->send_cnt += size >> 9;
  1431. return err;
  1432. }
  1433. static int _drbd_send_page(struct drbd_conf *mdev, struct page *page,
  1434. int offset, size_t size, unsigned msg_flags)
  1435. {
  1436. struct socket *socket = mdev->tconn->data.socket;
  1437. mm_segment_t oldfs = get_fs();
  1438. int len = size;
  1439. int err = -EIO;
  1440. /* e.g. XFS meta- & log-data is in slab pages, which have a
  1441. * page_count of 0 and/or have PageSlab() set.
  1442. * we cannot use send_page for those, as that does get_page();
  1443. * put_page(); and would cause either a VM_BUG directly, or
  1444. * __page_cache_release a page that would actually still be referenced
  1445. * by someone, leading to some obscure delayed Oops somewhere else. */
  1446. if (disable_sendpage || (page_count(page) < 1) || PageSlab(page))
  1447. return _drbd_no_send_page(mdev, page, offset, size, msg_flags);
  1448. msg_flags |= MSG_NOSIGNAL;
  1449. drbd_update_congested(mdev->tconn);
  1450. set_fs(KERNEL_DS);
  1451. do {
  1452. int sent;
  1453. sent = socket->ops->sendpage(socket, page, offset, len, msg_flags);
  1454. if (sent <= 0) {
  1455. if (sent == -EAGAIN) {
  1456. if (we_should_drop_the_connection(mdev->tconn, socket))
  1457. break;
  1458. continue;
  1459. }
  1460. dev_warn(DEV, "%s: size=%d len=%d sent=%d\n",
  1461. __func__, (int)size, len, sent);
  1462. if (sent < 0)
  1463. err = sent;
  1464. break;
  1465. }
  1466. len -= sent;
  1467. offset += sent;
  1468. } while (len > 0 /* THINK && mdev->cstate >= C_CONNECTED*/);
  1469. set_fs(oldfs);
  1470. clear_bit(NET_CONGESTED, &mdev->tconn->flags);
  1471. if (len == 0) {
  1472. err = 0;
  1473. mdev->send_cnt += size >> 9;
  1474. }
  1475. return err;
  1476. }
  1477. static int _drbd_send_bio(struct drbd_conf *mdev, struct bio *bio)
  1478. {
  1479. struct bio_vec *bvec;
  1480. int i;
  1481. /* hint all but last page with MSG_MORE */
  1482. bio_for_each_segment(bvec, bio, i) {
  1483. int err;
  1484. err = _drbd_no_send_page(mdev, bvec->bv_page,
  1485. bvec->bv_offset, bvec->bv_len,
  1486. i == bio->bi_vcnt - 1 ? 0 : MSG_MORE);
  1487. if (err)
  1488. return err;
  1489. }
  1490. return 0;
  1491. }
  1492. static int _drbd_send_zc_bio(struct drbd_conf *mdev, struct bio *bio)
  1493. {
  1494. struct bio_vec *bvec;
  1495. int i;
  1496. /* hint all but last page with MSG_MORE */
  1497. bio_for_each_segment(bvec, bio, i) {
  1498. int err;
  1499. err = _drbd_send_page(mdev, bvec->bv_page,
  1500. bvec->bv_offset, bvec->bv_len,
  1501. i == bio->bi_vcnt - 1 ? 0 : MSG_MORE);
  1502. if (err)
  1503. return err;
  1504. }
  1505. return 0;
  1506. }
  1507. static int _drbd_send_zc_ee(struct drbd_conf *mdev,
  1508. struct drbd_peer_request *peer_req)
  1509. {
  1510. struct page *page = peer_req->pages;
  1511. unsigned len = peer_req->i.size;
  1512. int err;
  1513. /* hint all but last page with MSG_MORE */
  1514. page_chain_for_each(page) {
  1515. unsigned l = min_t(unsigned, len, PAGE_SIZE);
  1516. err = _drbd_send_page(mdev, page, 0, l,
  1517. page_chain_next(page) ? MSG_MORE : 0);
  1518. if (err)
  1519. return err;
  1520. len -= l;
  1521. }
  1522. return 0;
  1523. }
  1524. static u32 bio_flags_to_wire(struct drbd_conf *mdev, unsigned long bi_rw)
  1525. {
  1526. if (mdev->tconn->agreed_pro_version >= 95)
  1527. return (bi_rw & REQ_SYNC ? DP_RW_SYNC : 0) |
  1528. (bi_rw & REQ_FUA ? DP_FUA : 0) |
  1529. (bi_rw & REQ_FLUSH ? DP_FLUSH : 0) |
  1530. (bi_rw & REQ_DISCARD ? DP_DISCARD : 0);
  1531. else
  1532. return bi_rw & REQ_SYNC ? DP_RW_SYNC : 0;
  1533. }
  1534. /* Used to send write requests
  1535. * R_PRIMARY -> Peer (P_DATA)
  1536. */
  1537. int drbd_send_dblock(struct drbd_conf *mdev, struct drbd_request *req)
  1538. {
  1539. struct drbd_socket *sock;
  1540. struct p_data *p;
  1541. unsigned int dp_flags = 0;
  1542. int dgs;
  1543. int err;
  1544. sock = &mdev->tconn->data;
  1545. p = drbd_prepare_command(mdev, sock);
  1546. dgs = mdev->tconn->integrity_tfm ? crypto_hash_digestsize(mdev->tconn->integrity_tfm) : 0;
  1547. if (!p)
  1548. return -EIO;
  1549. p->sector = cpu_to_be64(req->i.sector);
  1550. p->block_id = (unsigned long)req;
  1551. p->seq_num = cpu_to_be32(atomic_inc_return(&mdev->packet_seq));
  1552. dp_flags = bio_flags_to_wire(mdev, req->master_bio->bi_rw);
  1553. if (mdev->state.conn >= C_SYNC_SOURCE &&
  1554. mdev->state.conn <= C_PAUSED_SYNC_T)
  1555. dp_flags |= DP_MAY_SET_IN_SYNC;
  1556. if (mdev->tconn->agreed_pro_version >= 100) {
  1557. if (req->rq_state & RQ_EXP_RECEIVE_ACK)
  1558. dp_flags |= DP_SEND_RECEIVE_ACK;
  1559. if (req->rq_state & RQ_EXP_WRITE_ACK)
  1560. dp_flags |= DP_SEND_WRITE_ACK;
  1561. }
  1562. p->dp_flags = cpu_to_be32(dp_flags);
  1563. if (dgs)
  1564. drbd_csum_bio(mdev, mdev->tconn->integrity_tfm, req->master_bio, p + 1);
  1565. err = __send_command(mdev->tconn, mdev->vnr, sock, P_DATA, sizeof(*p) + dgs, NULL, req->i.size);
  1566. if (!err) {
  1567. /* For protocol A, we have to memcpy the payload into
  1568. * socket buffers, as we may complete right away
  1569. * as soon as we handed it over to tcp, at which point the data
  1570. * pages may become invalid.
  1571. *
  1572. * For data-integrity enabled, we copy it as well, so we can be
  1573. * sure that even if the bio pages may still be modified, it
  1574. * won't change the data on the wire, thus if the digest checks
  1575. * out ok after sending on this side, but does not fit on the
  1576. * receiving side, we sure have detected corruption elsewhere.
  1577. */
  1578. if (!(req->rq_state & (RQ_EXP_RECEIVE_ACK | RQ_EXP_WRITE_ACK)) || dgs)
  1579. err = _drbd_send_bio(mdev, req->master_bio);
  1580. else
  1581. err = _drbd_send_zc_bio(mdev, req->master_bio);
  1582. /* double check digest, sometimes buffers have been modified in flight. */
  1583. if (dgs > 0 && dgs <= 64) {
  1584. /* 64 byte, 512 bit, is the largest digest size
  1585. * currently supported in kernel crypto. */
  1586. unsigned char digest[64];
  1587. drbd_csum_bio(mdev, mdev->tconn->integrity_tfm, req->master_bio, digest);
  1588. if (memcmp(p + 1, digest, dgs)) {
  1589. dev_warn(DEV,
  1590. "Digest mismatch, buffer modified by upper layers during write: %llus +%u\n",
  1591. (unsigned long long)req->i.sector, req->i.size);
  1592. }
  1593. } /* else if (dgs > 64) {
  1594. ... Be noisy about digest too large ...
  1595. } */
  1596. }
  1597. mutex_unlock(&sock->mutex); /* locked by drbd_prepare_command() */
  1598. return err;
  1599. }
  1600. /* answer packet, used to send data back for read requests:
  1601. * Peer -> (diskless) R_PRIMARY (P_DATA_REPLY)
  1602. * C_SYNC_SOURCE -> C_SYNC_TARGET (P_RS_DATA_REPLY)
  1603. */
  1604. int drbd_send_block(struct drbd_conf *mdev, enum drbd_packet cmd,
  1605. struct drbd_peer_request *peer_req)
  1606. {
  1607. struct drbd_socket *sock;
  1608. struct p_data *p;
  1609. int err;
  1610. int dgs;
  1611. sock = &mdev->tconn->data;
  1612. p = drbd_prepare_command(mdev, sock);
  1613. dgs = mdev->tconn->integrity_tfm ? crypto_hash_digestsize(mdev->tconn->integrity_tfm) : 0;
  1614. if (!p)
  1615. return -EIO;
  1616. p->sector = cpu_to_be64(peer_req->i.sector);
  1617. p->block_id = peer_req->block_id;
  1618. p->seq_num = 0; /* unused */
  1619. p->dp_flags = 0;
  1620. if (dgs)
  1621. drbd_csum_ee(mdev, mdev->tconn->integrity_tfm, peer_req, p + 1);
  1622. err = __send_command(mdev->tconn, mdev->vnr, sock, cmd, sizeof(*p) + dgs, NULL, peer_req->i.size);
  1623. if (!err)
  1624. err = _drbd_send_zc_ee(mdev, peer_req);
  1625. mutex_unlock(&sock->mutex); /* locked by drbd_prepare_command() */
  1626. return err;
  1627. }
  1628. int drbd_send_out_of_sync(struct drbd_conf *mdev, struct drbd_request *req)
  1629. {
  1630. struct drbd_socket *sock;
  1631. struct p_block_desc *p;
  1632. sock = &mdev->tconn->data;
  1633. p = drbd_prepare_command(mdev, sock);
  1634. if (!p)
  1635. return -EIO;
  1636. p->sector = cpu_to_be64(req->i.sector);
  1637. p->blksize = cpu_to_be32(req->i.size);
  1638. return drbd_send_command(mdev, sock, P_OUT_OF_SYNC, sizeof(*p), NULL, 0);
  1639. }
  1640. /*
  1641. drbd_send distinguishes two cases:
  1642. Packets sent via the data socket "sock"
  1643. and packets sent via the meta data socket "msock"
  1644. sock msock
  1645. -----------------+-------------------------+------------------------------
  1646. timeout conf.timeout / 2 conf.timeout / 2
  1647. timeout action send a ping via msock Abort communication
  1648. and close all sockets
  1649. */
  1650. /*
  1651. * you must have down()ed the appropriate [m]sock_mutex elsewhere!
  1652. */
  1653. int drbd_send(struct drbd_tconn *tconn, struct socket *sock,
  1654. void *buf, size_t size, unsigned msg_flags)
  1655. {
  1656. struct kvec iov;
  1657. struct msghdr msg;
  1658. int rv, sent = 0;
  1659. if (!sock)
  1660. return -EBADR;
  1661. /* THINK if (signal_pending) return ... ? */
  1662. iov.iov_base = buf;
  1663. iov.iov_len = size;
  1664. msg.msg_name = NULL;
  1665. msg.msg_namelen = 0;
  1666. msg.msg_control = NULL;
  1667. msg.msg_controllen = 0;
  1668. msg.msg_flags = msg_flags | MSG_NOSIGNAL;
  1669. if (sock == tconn->data.socket) {
  1670. rcu_read_lock();
  1671. tconn->ko_count = rcu_dereference(tconn->net_conf)->ko_count;
  1672. rcu_read_unlock();
  1673. drbd_update_congested(tconn);
  1674. }
  1675. do {
  1676. /* STRANGE
  1677. * tcp_sendmsg does _not_ use its size parameter at all ?
  1678. *
  1679. * -EAGAIN on timeout, -EINTR on signal.
  1680. */
  1681. /* THINK
  1682. * do we need to block DRBD_SIG if sock == &meta.socket ??
  1683. * otherwise wake_asender() might interrupt some send_*Ack !
  1684. */
  1685. rv = kernel_sendmsg(sock, &msg, &iov, 1, size);
  1686. if (rv == -EAGAIN) {
  1687. if (we_should_drop_the_connection(tconn, sock))
  1688. break;
  1689. else
  1690. continue;
  1691. }
  1692. if (rv == -EINTR) {
  1693. flush_signals(current);
  1694. rv = 0;
  1695. }
  1696. if (rv < 0)
  1697. break;
  1698. sent += rv;
  1699. iov.iov_base += rv;
  1700. iov.iov_len -= rv;
  1701. } while (sent < size);
  1702. if (sock == tconn->data.socket)
  1703. clear_bit(NET_CONGESTED, &tconn->flags);
  1704. if (rv <= 0) {
  1705. if (rv != -EAGAIN) {
  1706. conn_err(tconn, "%s_sendmsg returned %d\n",
  1707. sock == tconn->meta.socket ? "msock" : "sock",
  1708. rv);
  1709. conn_request_state(tconn, NS(conn, C_BROKEN_PIPE), CS_HARD);
  1710. } else
  1711. conn_request_state(tconn, NS(conn, C_TIMEOUT), CS_HARD);
  1712. }
  1713. return sent;
  1714. }
  1715. /**
  1716. * drbd_send_all - Send an entire buffer
  1717. *
  1718. * Returns 0 upon success and a negative error value otherwise.
  1719. */
  1720. int drbd_send_all(struct drbd_tconn *tconn, struct socket *sock, void *buffer,
  1721. size_t size, unsigned msg_flags)
  1722. {
  1723. int err;
  1724. err = drbd_send(tconn, sock, buffer, size, msg_flags);
  1725. if (err < 0)
  1726. return err;
  1727. if (err != size)
  1728. return -EIO;
  1729. return 0;
  1730. }
  1731. static int drbd_open(struct block_device *bdev, fmode_t mode)
  1732. {
  1733. struct drbd_conf *mdev = bdev->bd_disk->private_data;
  1734. unsigned long flags;
  1735. int rv = 0;
  1736. mutex_lock(&drbd_main_mutex);
  1737. spin_lock_irqsave(&mdev->tconn->req_lock, flags);
  1738. /* to have a stable mdev->state.role
  1739. * and no race with updating open_cnt */
  1740. if (mdev->state.role != R_PRIMARY) {
  1741. if (mode & FMODE_WRITE)
  1742. rv = -EROFS;
  1743. else if (!allow_oos)
  1744. rv = -EMEDIUMTYPE;
  1745. }
  1746. if (!rv)
  1747. mdev->open_cnt++;
  1748. spin_unlock_irqrestore(&mdev->tconn->req_lock, flags);
  1749. mutex_unlock(&drbd_main_mutex);
  1750. return rv;
  1751. }
  1752. static int drbd_release(struct gendisk *gd, fmode_t mode)
  1753. {
  1754. struct drbd_conf *mdev = gd->private_data;
  1755. mutex_lock(&drbd_main_mutex);
  1756. mdev->open_cnt--;
  1757. mutex_unlock(&drbd_main_mutex);
  1758. return 0;
  1759. }
  1760. static void drbd_set_defaults(struct drbd_conf *mdev)
  1761. {
  1762. /* Beware! The actual layout differs
  1763. * between big endian and little endian */
  1764. mdev->state = (union drbd_dev_state) {
  1765. { .role = R_SECONDARY,
  1766. .peer = R_UNKNOWN,
  1767. .conn = C_STANDALONE,
  1768. .disk = D_DISKLESS,
  1769. .pdsk = D_UNKNOWN,
  1770. } };
  1771. }
  1772. void drbd_init_set_defaults(struct drbd_conf *mdev)
  1773. {
  1774. /* the memset(,0,) did most of this.
  1775. * note: only assignments, no allocation in here */
  1776. drbd_set_defaults(mdev);
  1777. atomic_set(&mdev->ap_bio_cnt, 0);
  1778. atomic_set(&mdev->ap_pending_cnt, 0);
  1779. atomic_set(&mdev->rs_pending_cnt, 0);
  1780. atomic_set(&mdev->unacked_cnt, 0);
  1781. atomic_set(&mdev->local_cnt, 0);
  1782. atomic_set(&mdev->pp_in_use_by_net, 0);
  1783. atomic_set(&mdev->rs_sect_in, 0);
  1784. atomic_set(&mdev->rs_sect_ev, 0);
  1785. atomic_set(&mdev->ap_in_flight, 0);
  1786. atomic_set(&mdev->md_io_in_use, 0);
  1787. mutex_init(&mdev->own_state_mutex);
  1788. mdev->state_mutex = &mdev->own_state_mutex;
  1789. spin_lock_init(&mdev->al_lock);
  1790. spin_lock_init(&mdev->peer_seq_lock);
  1791. INIT_LIST_HEAD(&mdev->active_ee);
  1792. INIT_LIST_HEAD(&mdev->sync_ee);
  1793. INIT_LIST_HEAD(&mdev->done_ee);
  1794. INIT_LIST_HEAD(&mdev->read_ee);
  1795. INIT_LIST_HEAD(&mdev->net_ee);
  1796. INIT_LIST_HEAD(&mdev->resync_reads);
  1797. INIT_LIST_HEAD(&mdev->resync_work.list);
  1798. INIT_LIST_HEAD(&mdev->unplug_work.list);
  1799. INIT_LIST_HEAD(&mdev->go_diskless.list);
  1800. INIT_LIST_HEAD(&mdev->md_sync_work.list);
  1801. INIT_LIST_HEAD(&mdev->start_resync_work.list);
  1802. INIT_LIST_HEAD(&mdev->bm_io_work.w.list);
  1803. mdev->resync_work.cb = w_resync_timer;
  1804. mdev->unplug_work.cb = w_send_write_hint;
  1805. mdev->go_diskless.cb = w_go_diskless;
  1806. mdev->md_sync_work.cb = w_md_sync;
  1807. mdev->bm_io_work.w.cb = w_bitmap_io;
  1808. mdev->start_resync_work.cb = w_start_resync;
  1809. mdev->resync_work.mdev = mdev;
  1810. mdev->unplug_work.mdev = mdev;
  1811. mdev->go_diskless.mdev = mdev;
  1812. mdev->md_sync_work.mdev = mdev;
  1813. mdev->bm_io_work.w.mdev = mdev;
  1814. mdev->start_resync_work.mdev = mdev;
  1815. init_timer(&mdev->resync_timer);
  1816. init_timer(&mdev->md_sync_timer);
  1817. init_timer(&mdev->start_resync_timer);
  1818. init_timer(&mdev->request_timer);
  1819. mdev->resync_timer.function = resync_timer_fn;
  1820. mdev->resync_timer.data = (unsigned long) mdev;
  1821. mdev->md_sync_timer.function = md_sync_timer_fn;
  1822. mdev->md_sync_timer.data = (unsigned long) mdev;
  1823. mdev->start_resync_timer.function = start_resync_timer_fn;
  1824. mdev->start_resync_timer.data = (unsigned long) mdev;
  1825. mdev->request_timer.function = request_timer_fn;
  1826. mdev->request_timer.data = (unsigned long) mdev;
  1827. init_waitqueue_head(&mdev->misc_wait);
  1828. init_waitqueue_head(&mdev->state_wait);
  1829. init_waitqueue_head(&mdev->ee_wait);
  1830. init_waitqueue_head(&mdev->al_wait);
  1831. init_waitqueue_head(&mdev->seq_wait);
  1832. mdev->resync_wenr = LC_FREE;
  1833. mdev->peer_max_bio_size = DRBD_MAX_BIO_SIZE_SAFE;
  1834. mdev->local_max_bio_size = DRBD_MAX_BIO_SIZE_SAFE;
  1835. }
  1836. void drbd_mdev_cleanup(struct drbd_conf *mdev)
  1837. {
  1838. int i;
  1839. if (mdev->tconn->receiver.t_state != NONE)
  1840. dev_err(DEV, "ASSERT FAILED: receiver t_state == %d expected 0.\n",
  1841. mdev->tconn->receiver.t_state);
  1842. mdev->al_writ_cnt =
  1843. mdev->bm_writ_cnt =
  1844. mdev->read_cnt =
  1845. mdev->recv_cnt =
  1846. mdev->send_cnt =
  1847. mdev->writ_cnt =
  1848. mdev->p_size =
  1849. mdev->rs_start =
  1850. mdev->rs_total =
  1851. mdev->rs_failed = 0;
  1852. mdev->rs_last_events = 0;
  1853. mdev->rs_last_sect_ev = 0;
  1854. for (i = 0; i < DRBD_SYNC_MARKS; i++) {
  1855. mdev->rs_mark_left[i] = 0;
  1856. mdev->rs_mark_time[i] = 0;
  1857. }
  1858. D_ASSERT(mdev->tconn->net_conf == NULL);
  1859. drbd_set_my_capacity(mdev, 0);
  1860. if (mdev->bitmap) {
  1861. /* maybe never allocated. */
  1862. drbd_bm_resize(mdev, 0, 1);
  1863. drbd_bm_cleanup(mdev);
  1864. }
  1865. drbd_free_bc(mdev->ldev);
  1866. mdev->ldev = NULL;
  1867. clear_bit(AL_SUSPENDED, &mdev->flags);
  1868. D_ASSERT(list_empty(&mdev->active_ee));
  1869. D_ASSERT(list_empty(&mdev->sync_ee));
  1870. D_ASSERT(list_empty(&mdev->done_ee));
  1871. D_ASSERT(list_empty(&mdev->read_ee));
  1872. D_ASSERT(list_empty(&mdev->net_ee));
  1873. D_ASSERT(list_empty(&mdev->resync_reads));
  1874. D_ASSERT(list_empty(&mdev->tconn->data.work.q));
  1875. D_ASSERT(list_empty(&mdev->tconn->meta.work.q));
  1876. D_ASSERT(list_empty(&mdev->resync_work.list));
  1877. D_ASSERT(list_empty(&mdev->unplug_work.list));
  1878. D_ASSERT(list_empty(&mdev->go_diskless.list));
  1879. drbd_set_defaults(mdev);
  1880. }
  1881. static void drbd_destroy_mempools(void)
  1882. {
  1883. struct page *page;
  1884. while (drbd_pp_pool) {
  1885. page = drbd_pp_pool;
  1886. drbd_pp_pool = (struct page *)page_private(page);
  1887. __free_page(page);
  1888. drbd_pp_vacant--;
  1889. }
  1890. /* D_ASSERT(atomic_read(&drbd_pp_vacant)==0); */
  1891. if (drbd_md_io_bio_set)
  1892. bioset_free(drbd_md_io_bio_set);
  1893. if (drbd_md_io_page_pool)
  1894. mempool_destroy(drbd_md_io_page_pool);
  1895. if (drbd_ee_mempool)
  1896. mempool_destroy(drbd_ee_mempool);
  1897. if (drbd_request_mempool)
  1898. mempool_destroy(drbd_request_mempool);
  1899. if (drbd_ee_cache)
  1900. kmem_cache_destroy(drbd_ee_cache);
  1901. if (drbd_request_cache)
  1902. kmem_cache_destroy(drbd_request_cache);
  1903. if (drbd_bm_ext_cache)
  1904. kmem_cache_destroy(drbd_bm_ext_cache);
  1905. if (drbd_al_ext_cache)
  1906. kmem_cache_destroy(drbd_al_ext_cache);
  1907. drbd_md_io_bio_set = NULL;
  1908. drbd_md_io_page_pool = NULL;
  1909. drbd_ee_mempool = NULL;
  1910. drbd_request_mempool = NULL;
  1911. drbd_ee_cache = NULL;
  1912. drbd_request_cache = NULL;
  1913. drbd_bm_ext_cache = NULL;
  1914. drbd_al_ext_cache = NULL;
  1915. return;
  1916. }
  1917. static int drbd_create_mempools(void)
  1918. {
  1919. struct page *page;
  1920. const int number = (DRBD_MAX_BIO_SIZE/PAGE_SIZE) * minor_count;
  1921. int i;
  1922. /* prepare our caches and mempools */
  1923. drbd_request_mempool = NULL;
  1924. drbd_ee_cache = NULL;
  1925. drbd_request_cache = NULL;
  1926. drbd_bm_ext_cache = NULL;
  1927. drbd_al_ext_cache = NULL;
  1928. drbd_pp_pool = NULL;
  1929. drbd_md_io_page_pool = NULL;
  1930. drbd_md_io_bio_set = NULL;
  1931. /* caches */
  1932. drbd_request_cache = kmem_cache_create(
  1933. "drbd_req", sizeof(struct drbd_request), 0, 0, NULL);
  1934. if (drbd_request_cache == NULL)
  1935. goto Enomem;
  1936. drbd_ee_cache = kmem_cache_create(
  1937. "drbd_ee", sizeof(struct drbd_peer_request), 0, 0, NULL);
  1938. if (drbd_ee_cache == NULL)
  1939. goto Enomem;
  1940. drbd_bm_ext_cache = kmem_cache_create(
  1941. "drbd_bm", sizeof(struct bm_extent), 0, 0, NULL);
  1942. if (drbd_bm_ext_cache == NULL)
  1943. goto Enomem;
  1944. drbd_al_ext_cache = kmem_cache_create(
  1945. "drbd_al", sizeof(struct lc_element), 0, 0, NULL);
  1946. if (drbd_al_ext_cache == NULL)
  1947. goto Enomem;
  1948. /* mempools */
  1949. drbd_md_io_bio_set = bioset_create(DRBD_MIN_POOL_PAGES, 0);
  1950. if (drbd_md_io_bio_set == NULL)
  1951. goto Enomem;
  1952. drbd_md_io_page_pool = mempool_create_page_pool(DRBD_MIN_POOL_PAGES, 0);
  1953. if (drbd_md_io_page_pool == NULL)
  1954. goto Enomem;
  1955. drbd_request_mempool = mempool_create(number,
  1956. mempool_alloc_slab, mempool_free_slab, drbd_request_cache);
  1957. if (drbd_request_mempool == NULL)
  1958. goto Enomem;
  1959. drbd_ee_mempool = mempool_create(number,
  1960. mempool_alloc_slab, mempool_free_slab, drbd_ee_cache);
  1961. if (drbd_ee_mempool == NULL)
  1962. goto Enomem;
  1963. /* drbd's page pool */
  1964. spin_lock_init(&drbd_pp_lock);
  1965. for (i = 0; i < number; i++) {
  1966. page = alloc_page(GFP_HIGHUSER);
  1967. if (!page)
  1968. goto Enomem;
  1969. set_page_private(page, (unsigned long)drbd_pp_pool);
  1970. drbd_pp_pool = page;
  1971. }
  1972. drbd_pp_vacant = number;
  1973. return 0;
  1974. Enomem:
  1975. drbd_destroy_mempools(); /* in case we allocated some */
  1976. return -ENOMEM;
  1977. }
  1978. static int drbd_notify_sys(struct notifier_block *this, unsigned long code,
  1979. void *unused)
  1980. {
  1981. /* just so we have it. you never know what interesting things we
  1982. * might want to do here some day...
  1983. */
  1984. return NOTIFY_DONE;
  1985. }
  1986. static struct notifier_block drbd_notifier = {
  1987. .notifier_call = drbd_notify_sys,
  1988. };
  1989. static void drbd_release_all_peer_reqs(struct drbd_conf *mdev)
  1990. {
  1991. int rr;
  1992. rr = drbd_free_peer_reqs(mdev, &mdev->active_ee);
  1993. if (rr)
  1994. dev_err(DEV, "%d EEs in active list found!\n", rr);
  1995. rr = drbd_free_peer_reqs(mdev, &mdev->sync_ee);
  1996. if (rr)
  1997. dev_err(DEV, "%d EEs in sync list found!\n", rr);
  1998. rr = drbd_free_peer_reqs(mdev, &mdev->read_ee);
  1999. if (rr)
  2000. dev_err(DEV, "%d EEs in read list found!\n", rr);
  2001. rr = drbd_free_peer_reqs(mdev, &mdev->done_ee);
  2002. if (rr)
  2003. dev_err(DEV, "%d EEs in done list found!\n", rr);
  2004. rr = drbd_free_peer_reqs(mdev, &mdev->net_ee);
  2005. if (rr)
  2006. dev_err(DEV, "%d EEs in net list found!\n", rr);
  2007. }
  2008. /* caution. no locking. */
  2009. void drbd_minor_destroy(struct kref *kref)
  2010. {
  2011. struct drbd_conf *mdev = container_of(kref, struct drbd_conf, kref);
  2012. struct drbd_tconn *tconn = mdev->tconn;
  2013. del_timer_sync(&mdev->request_timer);
  2014. /* paranoia asserts */
  2015. D_ASSERT(mdev->open_cnt == 0);
  2016. D_ASSERT(list_empty(&mdev->tconn->data.work.q));
  2017. /* end paranoia asserts */
  2018. /* cleanup stuff that may have been allocated during
  2019. * device (re-)configuration or state changes */
  2020. if (mdev->this_bdev)
  2021. bdput(mdev->this_bdev);
  2022. drbd_free_bc(mdev->ldev);
  2023. mdev->ldev = NULL;
  2024. drbd_release_all_peer_reqs(mdev);
  2025. lc_destroy(mdev->act_log);
  2026. lc_destroy(mdev->resync);
  2027. kfree(mdev->p_uuid);
  2028. /* mdev->p_uuid = NULL; */
  2029. if (mdev->bitmap) /* should no longer be there. */
  2030. drbd_bm_cleanup(mdev);
  2031. __free_page(mdev->md_io_page);
  2032. put_disk(mdev->vdisk);
  2033. blk_cleanup_queue(mdev->rq_queue);
  2034. kfree(mdev->rs_plan_s);
  2035. kfree(mdev);
  2036. kref_put(&tconn->kref, &conn_destroy);
  2037. }
  2038. /* One global retry thread, if we need to push back some bio and have it
  2039. * reinserted through our make request function.
  2040. */
  2041. static struct retry_worker {
  2042. struct workqueue_struct *wq;
  2043. struct work_struct worker;
  2044. spinlock_t lock;
  2045. struct list_head writes;
  2046. } retry;
  2047. static void do_retry(struct work_struct *ws)
  2048. {
  2049. struct retry_worker *retry = container_of(ws, struct retry_worker, worker);
  2050. LIST_HEAD(writes);
  2051. struct drbd_request *req, *tmp;
  2052. spin_lock_irq(&retry->lock);
  2053. list_splice_init(&retry->writes, &writes);
  2054. spin_unlock_irq(&retry->lock);
  2055. list_for_each_entry_safe(req, tmp, &writes, tl_requests) {
  2056. struct drbd_conf *mdev = req->w.mdev;
  2057. struct bio *bio = req->master_bio;
  2058. unsigned long start_time = req->start_time;
  2059. /* We have exclusive access to this request object.
  2060. * If it had not been RQ_POSTPONED, the code path which queued
  2061. * it here would have completed and freed it already.
  2062. */
  2063. mempool_free(req, drbd_request_mempool);
  2064. /* A single suspended or otherwise blocking device may stall
  2065. * all others as well. Fortunately, this code path is to
  2066. * recover from a situation that "should not happen":
  2067. * concurrent writes in multi-primary setup.
  2068. * In a "normal" lifecycle, this workqueue is supposed to be
  2069. * destroyed without ever doing anything.
  2070. * If it turns out to be an issue anyways, we can do per
  2071. * resource (replication group) or per device (minor) retry
  2072. * workqueues instead.
  2073. */
  2074. /* We are not just doing generic_make_request(),
  2075. * as we want to keep the start_time information. */
  2076. do {
  2077. inc_ap_bio(mdev);
  2078. } while(__drbd_make_request(mdev, bio, start_time));
  2079. }
  2080. }
  2081. void drbd_restart_request(struct drbd_request *req)
  2082. {
  2083. unsigned long flags;
  2084. spin_lock_irqsave(&retry.lock, flags);
  2085. list_move_tail(&req->tl_requests, &retry.writes);
  2086. spin_unlock_irqrestore(&retry.lock, flags);
  2087. /* Drop the extra reference that would otherwise
  2088. * have been dropped by complete_master_bio.
  2089. * do_retry() needs to grab a new one. */
  2090. dec_ap_bio(req->w.mdev);
  2091. queue_work(retry.wq, &retry.worker);
  2092. }
  2093. static void drbd_cleanup(void)
  2094. {
  2095. unsigned int i;
  2096. struct drbd_conf *mdev;
  2097. struct drbd_tconn *tconn, *tmp;
  2098. unregister_reboot_notifier(&drbd_notifier);
  2099. /* first remove proc,
  2100. * drbdsetup uses it's presence to detect
  2101. * whether DRBD is loaded.
  2102. * If we would get stuck in proc removal,
  2103. * but have netlink already deregistered,
  2104. * some drbdsetup commands may wait forever
  2105. * for an answer.
  2106. */
  2107. if (drbd_proc)
  2108. remove_proc_entry("drbd", NULL);
  2109. if (retry.wq)
  2110. destroy_workqueue(retry.wq);
  2111. drbd_genl_unregister();
  2112. idr_for_each_entry(&minors, mdev, i) {
  2113. idr_remove(&minors, mdev_to_minor(mdev));
  2114. idr_remove(&mdev->tconn->volumes, mdev->vnr);
  2115. del_gendisk(mdev->vdisk);
  2116. /* synchronize_rcu(); No other threads running at this point */
  2117. kref_put(&mdev->kref, &drbd_minor_destroy);
  2118. }
  2119. /* not _rcu since, no other updater anymore. Genl already unregistered */
  2120. list_for_each_entry_safe(tconn, tmp, &drbd_tconns, all_tconn) {
  2121. list_del(&tconn->all_tconn); /* not _rcu no proc, not other threads */
  2122. /* synchronize_rcu(); */
  2123. kref_put(&tconn->kref, &conn_destroy);
  2124. }
  2125. drbd_destroy_mempools();
  2126. unregister_blkdev(DRBD_MAJOR, "drbd");
  2127. idr_destroy(&minors);
  2128. printk(KERN_INFO "drbd: module cleanup done.\n");
  2129. }
  2130. /**
  2131. * drbd_congested() - Callback for pdflush
  2132. * @congested_data: User data
  2133. * @bdi_bits: Bits pdflush is currently interested in
  2134. *
  2135. * Returns 1<<BDI_async_congested and/or 1<<BDI_sync_congested if we are congested.
  2136. */
  2137. static int drbd_congested(void *congested_data, int bdi_bits)
  2138. {
  2139. struct drbd_conf *mdev = congested_data;
  2140. struct request_queue *q;
  2141. char reason = '-';
  2142. int r = 0;
  2143. if (!may_inc_ap_bio(mdev)) {
  2144. /* DRBD has frozen IO */
  2145. r = bdi_bits;
  2146. reason = 'd';
  2147. goto out;
  2148. }
  2149. if (get_ldev(mdev)) {
  2150. q = bdev_get_queue(mdev->ldev->backing_bdev);
  2151. r = bdi_congested(&q->backing_dev_info, bdi_bits);
  2152. put_ldev(mdev);
  2153. if (r)
  2154. reason = 'b';
  2155. }
  2156. if (bdi_bits & (1 << BDI_async_congested) && test_bit(NET_CONGESTED, &mdev->tconn->flags)) {
  2157. r |= (1 << BDI_async_congested);
  2158. reason = reason == 'b' ? 'a' : 'n';
  2159. }
  2160. out:
  2161. mdev->congestion_reason = reason;
  2162. return r;
  2163. }
  2164. static void drbd_init_workqueue(struct drbd_work_queue* wq)
  2165. {
  2166. sema_init(&wq->s, 0);
  2167. spin_lock_init(&wq->q_lock);
  2168. INIT_LIST_HEAD(&wq->q);
  2169. }
  2170. struct drbd_tconn *conn_get_by_name(const char *name)
  2171. {
  2172. struct drbd_tconn *tconn;
  2173. if (!name || !name[0])
  2174. return NULL;
  2175. rcu_read_lock();
  2176. list_for_each_entry_rcu(tconn, &drbd_tconns, all_tconn) {
  2177. if (!strcmp(tconn->name, name)) {
  2178. kref_get(&tconn->kref);
  2179. goto found;
  2180. }
  2181. }
  2182. tconn = NULL;
  2183. found:
  2184. rcu_read_unlock();
  2185. return tconn;
  2186. }
  2187. struct drbd_tconn *conn_get_by_addrs(void *my_addr, int my_addr_len,
  2188. void *peer_addr, int peer_addr_len)
  2189. {
  2190. struct drbd_tconn *tconn;
  2191. rcu_read_lock();
  2192. list_for_each_entry_rcu(tconn, &drbd_tconns, all_tconn) {
  2193. if (tconn->my_addr_len == my_addr_len &&
  2194. tconn->peer_addr_len == peer_addr_len &&
  2195. !memcmp(&tconn->my_addr, my_addr, my_addr_len) &&
  2196. !memcmp(&tconn->peer_addr, peer_addr, peer_addr_len)) {
  2197. kref_get(&tconn->kref);
  2198. goto found;
  2199. }
  2200. }
  2201. tconn = NULL;
  2202. found:
  2203. rcu_read_unlock();
  2204. return tconn;
  2205. }
  2206. static int drbd_alloc_socket(struct drbd_socket *socket)
  2207. {
  2208. socket->rbuf = (void *) __get_free_page(GFP_KERNEL);
  2209. if (!socket->rbuf)
  2210. return -ENOMEM;
  2211. socket->sbuf = (void *) __get_free_page(GFP_KERNEL);
  2212. if (!socket->sbuf)
  2213. return -ENOMEM;
  2214. return 0;
  2215. }
  2216. static void drbd_free_socket(struct drbd_socket *socket)
  2217. {
  2218. free_page((unsigned long) socket->sbuf);
  2219. free_page((unsigned long) socket->rbuf);
  2220. }
  2221. void conn_free_crypto(struct drbd_tconn *tconn)
  2222. {
  2223. drbd_free_sock(tconn);
  2224. crypto_free_hash(tconn->csums_tfm);
  2225. crypto_free_hash(tconn->verify_tfm);
  2226. crypto_free_hash(tconn->cram_hmac_tfm);
  2227. crypto_free_hash(tconn->integrity_tfm);
  2228. crypto_free_hash(tconn->peer_integrity_tfm);
  2229. kfree(tconn->int_dig_in);
  2230. kfree(tconn->int_dig_vv);
  2231. tconn->csums_tfm = NULL;
  2232. tconn->verify_tfm = NULL;
  2233. tconn->cram_hmac_tfm = NULL;
  2234. tconn->integrity_tfm = NULL;
  2235. tconn->peer_integrity_tfm = NULL;
  2236. tconn->int_dig_in = NULL;
  2237. tconn->int_dig_vv = NULL;
  2238. }
  2239. int set_resource_options(struct drbd_tconn *tconn, struct res_opts *res_opts)
  2240. {
  2241. cpumask_var_t new_cpu_mask;
  2242. int err;
  2243. if (!zalloc_cpumask_var(&new_cpu_mask, GFP_KERNEL))
  2244. return -ENOMEM;
  2245. /*
  2246. retcode = ERR_NOMEM;
  2247. drbd_msg_put_info("unable to allocate cpumask");
  2248. */
  2249. /* silently ignore cpu mask on UP kernel */
  2250. if (nr_cpu_ids > 1 && res_opts->cpu_mask[0] != 0) {
  2251. /* FIXME: Get rid of constant 32 here */
  2252. err = bitmap_parse(res_opts->cpu_mask, 32,
  2253. cpumask_bits(new_cpu_mask), nr_cpu_ids);
  2254. if (err) {
  2255. conn_warn(tconn, "bitmap_parse() failed with %d\n", err);
  2256. /* retcode = ERR_CPU_MASK_PARSE; */
  2257. goto fail;
  2258. }
  2259. }
  2260. tconn->res_opts = *res_opts;
  2261. if (!cpumask_equal(tconn->cpu_mask, new_cpu_mask)) {
  2262. cpumask_copy(tconn->cpu_mask, new_cpu_mask);
  2263. drbd_calc_cpu_mask(tconn);
  2264. tconn->receiver.reset_cpu_mask = 1;
  2265. tconn->asender.reset_cpu_mask = 1;
  2266. tconn->worker.reset_cpu_mask = 1;
  2267. }
  2268. err = 0;
  2269. fail:
  2270. free_cpumask_var(new_cpu_mask);
  2271. return err;
  2272. }
  2273. /* caller must be under genl_lock() */
  2274. struct drbd_tconn *conn_create(const char *name, struct res_opts *res_opts)
  2275. {
  2276. struct drbd_tconn *tconn;
  2277. tconn = kzalloc(sizeof(struct drbd_tconn), GFP_KERNEL);
  2278. if (!tconn)
  2279. return NULL;
  2280. tconn->name = kstrdup(name, GFP_KERNEL);
  2281. if (!tconn->name)
  2282. goto fail;
  2283. if (drbd_alloc_socket(&tconn->data))
  2284. goto fail;
  2285. if (drbd_alloc_socket(&tconn->meta))
  2286. goto fail;
  2287. if (!zalloc_cpumask_var(&tconn->cpu_mask, GFP_KERNEL))
  2288. goto fail;
  2289. if (set_resource_options(tconn, res_opts))
  2290. goto fail;
  2291. if (!tl_init(tconn))
  2292. goto fail;
  2293. tconn->current_epoch = kzalloc(sizeof(struct drbd_epoch), GFP_KERNEL);
  2294. if (!tconn->current_epoch)
  2295. goto fail;
  2296. INIT_LIST_HEAD(&tconn->current_epoch->list);
  2297. tconn->epochs = 1;
  2298. spin_lock_init(&tconn->epoch_lock);
  2299. tconn->write_ordering = WO_bdev_flush;
  2300. tconn->cstate = C_STANDALONE;
  2301. mutex_init(&tconn->cstate_mutex);
  2302. spin_lock_init(&tconn->req_lock);
  2303. mutex_init(&tconn->conf_update);
  2304. init_waitqueue_head(&tconn->ping_wait);
  2305. idr_init(&tconn->volumes);
  2306. drbd_init_workqueue(&tconn->data.work);
  2307. mutex_init(&tconn->data.mutex);
  2308. drbd_init_workqueue(&tconn->meta.work);
  2309. mutex_init(&tconn->meta.mutex);
  2310. drbd_thread_init(tconn, &tconn->receiver, drbdd_init, "receiver");
  2311. drbd_thread_init(tconn, &tconn->worker, drbd_worker, "worker");
  2312. drbd_thread_init(tconn, &tconn->asender, drbd_asender, "asender");
  2313. kref_init(&tconn->kref);
  2314. list_add_tail_rcu(&tconn->all_tconn, &drbd_tconns);
  2315. return tconn;
  2316. fail:
  2317. kfree(tconn->current_epoch);
  2318. tl_cleanup(tconn);
  2319. free_cpumask_var(tconn->cpu_mask);
  2320. drbd_free_socket(&tconn->meta);
  2321. drbd_free_socket(&tconn->data);
  2322. kfree(tconn->name);
  2323. kfree(tconn);
  2324. return NULL;
  2325. }
  2326. void conn_destroy(struct kref *kref)
  2327. {
  2328. struct drbd_tconn *tconn = container_of(kref, struct drbd_tconn, kref);
  2329. if (atomic_read(&tconn->current_epoch->epoch_size) != 0)
  2330. conn_err(tconn, "epoch_size:%d\n", atomic_read(&tconn->current_epoch->epoch_size));
  2331. kfree(tconn->current_epoch);
  2332. idr_destroy(&tconn->volumes);
  2333. free_cpumask_var(tconn->cpu_mask);
  2334. drbd_free_socket(&tconn->meta);
  2335. drbd_free_socket(&tconn->data);
  2336. kfree(tconn->name);
  2337. kfree(tconn->int_dig_in);
  2338. kfree(tconn->int_dig_vv);
  2339. kfree(tconn);
  2340. }
  2341. enum drbd_ret_code conn_new_minor(struct drbd_tconn *tconn, unsigned int minor, int vnr)
  2342. {
  2343. struct drbd_conf *mdev;
  2344. struct gendisk *disk;
  2345. struct request_queue *q;
  2346. int vnr_got = vnr;
  2347. int minor_got = minor;
  2348. enum drbd_ret_code err = ERR_NOMEM;
  2349. mdev = minor_to_mdev(minor);
  2350. if (mdev)
  2351. return ERR_MINOR_EXISTS;
  2352. /* GFP_KERNEL, we are outside of all write-out paths */
  2353. mdev = kzalloc(sizeof(struct drbd_conf), GFP_KERNEL);
  2354. if (!mdev)
  2355. return ERR_NOMEM;
  2356. kref_get(&tconn->kref);
  2357. mdev->tconn = tconn;
  2358. mdev->minor = minor;
  2359. mdev->vnr = vnr;
  2360. drbd_init_set_defaults(mdev);
  2361. q = blk_alloc_queue(GFP_KERNEL);
  2362. if (!q)
  2363. goto out_no_q;
  2364. mdev->rq_queue = q;
  2365. q->queuedata = mdev;
  2366. disk = alloc_disk(1);
  2367. if (!disk)
  2368. goto out_no_disk;
  2369. mdev->vdisk = disk;
  2370. set_disk_ro(disk, true);
  2371. disk->queue = q;
  2372. disk->major = DRBD_MAJOR;
  2373. disk->first_minor = minor;
  2374. disk->fops = &drbd_ops;
  2375. sprintf(disk->disk_name, "drbd%d", minor);
  2376. disk->private_data = mdev;
  2377. mdev->this_bdev = bdget(MKDEV(DRBD_MAJOR, minor));
  2378. /* we have no partitions. we contain only ourselves. */
  2379. mdev->this_bdev->bd_contains = mdev->this_bdev;
  2380. q->backing_dev_info.congested_fn = drbd_congested;
  2381. q->backing_dev_info.congested_data = mdev;
  2382. blk_queue_make_request(q, drbd_make_request);
  2383. /* Setting the max_hw_sectors to an odd value of 8kibyte here
  2384. This triggers a max_bio_size message upon first attach or connect */
  2385. blk_queue_max_hw_sectors(q, DRBD_MAX_BIO_SIZE_SAFE >> 8);
  2386. blk_queue_bounce_limit(q, BLK_BOUNCE_ANY);
  2387. blk_queue_merge_bvec(q, drbd_merge_bvec);
  2388. q->queue_lock = &mdev->tconn->req_lock; /* needed since we use */
  2389. mdev->md_io_page = alloc_page(GFP_KERNEL);
  2390. if (!mdev->md_io_page)
  2391. goto out_no_io_page;
  2392. if (drbd_bm_init(mdev))
  2393. goto out_no_bitmap;
  2394. mdev->read_requests = RB_ROOT;
  2395. mdev->write_requests = RB_ROOT;
  2396. if (!idr_pre_get(&minors, GFP_KERNEL))
  2397. goto out_no_minor_idr;
  2398. if (idr_get_new_above(&minors, mdev, minor, &minor_got))
  2399. goto out_no_minor_idr;
  2400. if (minor_got != minor) {
  2401. err = ERR_MINOR_EXISTS;
  2402. drbd_msg_put_info("requested minor exists already");
  2403. goto out_idr_remove_minor;
  2404. }
  2405. if (!idr_pre_get(&tconn->volumes, GFP_KERNEL))
  2406. goto out_idr_remove_minor;
  2407. if (idr_get_new_above(&tconn->volumes, mdev, vnr, &vnr_got))
  2408. goto out_idr_remove_minor;
  2409. if (vnr_got != vnr) {
  2410. err = ERR_INVALID_REQUEST;
  2411. drbd_msg_put_info("requested volume exists already");
  2412. goto out_idr_remove_vol;
  2413. }
  2414. add_disk(disk);
  2415. kref_init(&mdev->kref); /* one ref for both idrs and the the add_disk */
  2416. /* inherit the connection state */
  2417. mdev->state.conn = tconn->cstate;
  2418. if (mdev->state.conn == C_WF_REPORT_PARAMS)
  2419. drbd_connected(mdev);
  2420. return NO_ERROR;
  2421. out_idr_remove_vol:
  2422. idr_remove(&tconn->volumes, vnr_got);
  2423. out_idr_remove_minor:
  2424. idr_remove(&minors, minor_got);
  2425. synchronize_rcu();
  2426. out_no_minor_idr:
  2427. drbd_bm_cleanup(mdev);
  2428. out_no_bitmap:
  2429. __free_page(mdev->md_io_page);
  2430. out_no_io_page:
  2431. put_disk(disk);
  2432. out_no_disk:
  2433. blk_cleanup_queue(q);
  2434. out_no_q:
  2435. kfree(mdev);
  2436. kref_put(&tconn->kref, &conn_destroy);
  2437. return err;
  2438. }
  2439. int __init drbd_init(void)
  2440. {
  2441. int err;
  2442. if (minor_count < DRBD_MINOR_COUNT_MIN || minor_count > DRBD_MINOR_COUNT_MAX) {
  2443. printk(KERN_ERR
  2444. "drbd: invalid minor_count (%d)\n", minor_count);
  2445. #ifdef MODULE
  2446. return -EINVAL;
  2447. #else
  2448. minor_count = DRBD_MINOR_COUNT_DEF;
  2449. #endif
  2450. }
  2451. err = register_blkdev(DRBD_MAJOR, "drbd");
  2452. if (err) {
  2453. printk(KERN_ERR
  2454. "drbd: unable to register block device major %d\n",
  2455. DRBD_MAJOR);
  2456. return err;
  2457. }
  2458. err = drbd_genl_register();
  2459. if (err) {
  2460. printk(KERN_ERR "drbd: unable to register generic netlink family\n");
  2461. goto fail;
  2462. }
  2463. register_reboot_notifier(&drbd_notifier);
  2464. /*
  2465. * allocate all necessary structs
  2466. */
  2467. err = -ENOMEM;
  2468. init_waitqueue_head(&drbd_pp_wait);
  2469. drbd_proc = NULL; /* play safe for drbd_cleanup */
  2470. idr_init(&minors);
  2471. err = drbd_create_mempools();
  2472. if (err)
  2473. goto fail;
  2474. drbd_proc = proc_create_data("drbd", S_IFREG | S_IRUGO , NULL, &drbd_proc_fops, NULL);
  2475. if (!drbd_proc) {
  2476. printk(KERN_ERR "drbd: unable to register proc file\n");
  2477. goto fail;
  2478. }
  2479. rwlock_init(&global_state_lock);
  2480. INIT_LIST_HEAD(&drbd_tconns);
  2481. retry.wq = create_singlethread_workqueue("drbd-reissue");
  2482. if (!retry.wq) {
  2483. printk(KERN_ERR "drbd: unable to create retry workqueue\n");
  2484. goto fail;
  2485. }
  2486. INIT_WORK(&retry.worker, do_retry);
  2487. spin_lock_init(&retry.lock);
  2488. INIT_LIST_HEAD(&retry.writes);
  2489. printk(KERN_INFO "drbd: initialized. "
  2490. "Version: " REL_VERSION " (api:%d/proto:%d-%d)\n",
  2491. API_VERSION, PRO_VERSION_MIN, PRO_VERSION_MAX);
  2492. printk(KERN_INFO "drbd: %s\n", drbd_buildtag());
  2493. printk(KERN_INFO "drbd: registered as block device major %d\n",
  2494. DRBD_MAJOR);
  2495. return 0; /* Success! */
  2496. fail:
  2497. drbd_cleanup();
  2498. if (err == -ENOMEM)
  2499. /* currently always the case */
  2500. printk(KERN_ERR "drbd: ran out of memory\n");
  2501. else
  2502. printk(KERN_ERR "drbd: initialization failure\n");
  2503. return err;
  2504. }
  2505. void drbd_free_bc(struct drbd_backing_dev *ldev)
  2506. {
  2507. if (ldev == NULL)
  2508. return;
  2509. blkdev_put(ldev->backing_bdev, FMODE_READ | FMODE_WRITE | FMODE_EXCL);
  2510. blkdev_put(ldev->md_bdev, FMODE_READ | FMODE_WRITE | FMODE_EXCL);
  2511. kfree(ldev);
  2512. }
  2513. void drbd_free_sock(struct drbd_tconn *tconn)
  2514. {
  2515. if (tconn->data.socket) {
  2516. mutex_lock(&tconn->data.mutex);
  2517. kernel_sock_shutdown(tconn->data.socket, SHUT_RDWR);
  2518. sock_release(tconn->data.socket);
  2519. tconn->data.socket = NULL;
  2520. mutex_unlock(&tconn->data.mutex);
  2521. }
  2522. if (tconn->meta.socket) {
  2523. mutex_lock(&tconn->meta.mutex);
  2524. kernel_sock_shutdown(tconn->meta.socket, SHUT_RDWR);
  2525. sock_release(tconn->meta.socket);
  2526. tconn->meta.socket = NULL;
  2527. mutex_unlock(&tconn->meta.mutex);
  2528. }
  2529. }
  2530. /* meta data management */
  2531. struct meta_data_on_disk {
  2532. u64 la_size; /* last agreed size. */
  2533. u64 uuid[UI_SIZE]; /* UUIDs. */
  2534. u64 device_uuid;
  2535. u64 reserved_u64_1;
  2536. u32 flags; /* MDF */
  2537. u32 magic;
  2538. u32 md_size_sect;
  2539. u32 al_offset; /* offset to this block */
  2540. u32 al_nr_extents; /* important for restoring the AL */
  2541. /* `-- act_log->nr_elements <-- ldev->dc.al_extents */
  2542. u32 bm_offset; /* offset to the bitmap, from here */
  2543. u32 bm_bytes_per_bit; /* BM_BLOCK_SIZE */
  2544. u32 la_peer_max_bio_size; /* last peer max_bio_size */
  2545. u32 reserved_u32[3];
  2546. } __packed;
  2547. /**
  2548. * drbd_md_sync() - Writes the meta data super block if the MD_DIRTY flag bit is set
  2549. * @mdev: DRBD device.
  2550. */
  2551. void drbd_md_sync(struct drbd_conf *mdev)
  2552. {
  2553. struct meta_data_on_disk *buffer;
  2554. sector_t sector;
  2555. int i;
  2556. del_timer(&mdev->md_sync_timer);
  2557. /* timer may be rearmed by drbd_md_mark_dirty() now. */
  2558. if (!test_and_clear_bit(MD_DIRTY, &mdev->flags))
  2559. return;
  2560. /* We use here D_FAILED and not D_ATTACHING because we try to write
  2561. * metadata even if we detach due to a disk failure! */
  2562. if (!get_ldev_if_state(mdev, D_FAILED))
  2563. return;
  2564. buffer = drbd_md_get_buffer(mdev);
  2565. if (!buffer)
  2566. goto out;
  2567. memset(buffer, 0, 512);
  2568. buffer->la_size = cpu_to_be64(drbd_get_capacity(mdev->this_bdev));
  2569. for (i = UI_CURRENT; i < UI_SIZE; i++)
  2570. buffer->uuid[i] = cpu_to_be64(mdev->ldev->md.uuid[i]);
  2571. buffer->flags = cpu_to_be32(mdev->ldev->md.flags);
  2572. buffer->magic = cpu_to_be32(DRBD_MD_MAGIC_84_UNCLEAN);
  2573. buffer->md_size_sect = cpu_to_be32(mdev->ldev->md.md_size_sect);
  2574. buffer->al_offset = cpu_to_be32(mdev->ldev->md.al_offset);
  2575. buffer->al_nr_extents = cpu_to_be32(mdev->act_log->nr_elements);
  2576. buffer->bm_bytes_per_bit = cpu_to_be32(BM_BLOCK_SIZE);
  2577. buffer->device_uuid = cpu_to_be64(mdev->ldev->md.device_uuid);
  2578. buffer->bm_offset = cpu_to_be32(mdev->ldev->md.bm_offset);
  2579. buffer->la_peer_max_bio_size = cpu_to_be32(mdev->peer_max_bio_size);
  2580. D_ASSERT(drbd_md_ss__(mdev, mdev->ldev) == mdev->ldev->md.md_offset);
  2581. sector = mdev->ldev->md.md_offset;
  2582. if (drbd_md_sync_page_io(mdev, mdev->ldev, sector, WRITE)) {
  2583. /* this was a try anyways ... */
  2584. dev_err(DEV, "meta data update failed!\n");
  2585. drbd_chk_io_error(mdev, 1, true);
  2586. }
  2587. /* Update mdev->ldev->md.la_size_sect,
  2588. * since we updated it on metadata. */
  2589. mdev->ldev->md.la_size_sect = drbd_get_capacity(mdev->this_bdev);
  2590. drbd_md_put_buffer(mdev);
  2591. out:
  2592. put_ldev(mdev);
  2593. }
  2594. /**
  2595. * drbd_md_read() - Reads in the meta data super block
  2596. * @mdev: DRBD device.
  2597. * @bdev: Device from which the meta data should be read in.
  2598. *
  2599. * Return 0 (NO_ERROR) on success, and an enum drbd_ret_code in case
  2600. * something goes wrong.
  2601. */
  2602. int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev)
  2603. {
  2604. struct meta_data_on_disk *buffer;
  2605. u32 magic, flags;
  2606. int i, rv = NO_ERROR;
  2607. if (!get_ldev_if_state(mdev, D_ATTACHING))
  2608. return ERR_IO_MD_DISK;
  2609. buffer = drbd_md_get_buffer(mdev);
  2610. if (!buffer)
  2611. goto out;
  2612. if (drbd_md_sync_page_io(mdev, bdev, bdev->md.md_offset, READ)) {
  2613. /* NOTE: can't do normal error processing here as this is
  2614. called BEFORE disk is attached */
  2615. dev_err(DEV, "Error while reading metadata.\n");
  2616. rv = ERR_IO_MD_DISK;
  2617. goto err;
  2618. }
  2619. magic = be32_to_cpu(buffer->magic);
  2620. flags = be32_to_cpu(buffer->flags);
  2621. if (magic == DRBD_MD_MAGIC_84_UNCLEAN ||
  2622. (magic == DRBD_MD_MAGIC_08 && !(flags & MDF_AL_CLEAN))) {
  2623. /* btw: that's Activity Log clean, not "all" clean. */
  2624. dev_err(DEV, "Found unclean meta data. Did you \"drbdadm apply-al\"?\n");
  2625. rv = ERR_MD_UNCLEAN;
  2626. goto err;
  2627. }
  2628. if (magic != DRBD_MD_MAGIC_08) {
  2629. if (magic == DRBD_MD_MAGIC_07)
  2630. dev_err(DEV, "Found old (0.7) meta data magic. Did you \"drbdadm create-md\"?\n");
  2631. else
  2632. dev_err(DEV, "Meta data magic not found. Did you \"drbdadm create-md\"?\n");
  2633. rv = ERR_MD_INVALID;
  2634. goto err;
  2635. }
  2636. if (be32_to_cpu(buffer->al_offset) != bdev->md.al_offset) {
  2637. dev_err(DEV, "unexpected al_offset: %d (expected %d)\n",
  2638. be32_to_cpu(buffer->al_offset), bdev->md.al_offset);
  2639. rv = ERR_MD_INVALID;
  2640. goto err;
  2641. }
  2642. if (be32_to_cpu(buffer->bm_offset) != bdev->md.bm_offset) {
  2643. dev_err(DEV, "unexpected bm_offset: %d (expected %d)\n",
  2644. be32_to_cpu(buffer->bm_offset), bdev->md.bm_offset);
  2645. rv = ERR_MD_INVALID;
  2646. goto err;
  2647. }
  2648. if (be32_to_cpu(buffer->md_size_sect) != bdev->md.md_size_sect) {
  2649. dev_err(DEV, "unexpected md_size: %u (expected %u)\n",
  2650. be32_to_cpu(buffer->md_size_sect), bdev->md.md_size_sect);
  2651. rv = ERR_MD_INVALID;
  2652. goto err;
  2653. }
  2654. if (be32_to_cpu(buffer->bm_bytes_per_bit) != BM_BLOCK_SIZE) {
  2655. dev_err(DEV, "unexpected bm_bytes_per_bit: %u (expected %u)\n",
  2656. be32_to_cpu(buffer->bm_bytes_per_bit), BM_BLOCK_SIZE);
  2657. rv = ERR_MD_INVALID;
  2658. goto err;
  2659. }
  2660. bdev->md.la_size_sect = be64_to_cpu(buffer->la_size);
  2661. for (i = UI_CURRENT; i < UI_SIZE; i++)
  2662. bdev->md.uuid[i] = be64_to_cpu(buffer->uuid[i]);
  2663. bdev->md.flags = be32_to_cpu(buffer->flags);
  2664. bdev->md.device_uuid = be64_to_cpu(buffer->device_uuid);
  2665. spin_lock_irq(&mdev->tconn->req_lock);
  2666. if (mdev->state.conn < C_CONNECTED) {
  2667. int peer;
  2668. peer = be32_to_cpu(buffer->la_peer_max_bio_size);
  2669. peer = max_t(int, peer, DRBD_MAX_BIO_SIZE_SAFE);
  2670. mdev->peer_max_bio_size = peer;
  2671. }
  2672. spin_unlock_irq(&mdev->tconn->req_lock);
  2673. err:
  2674. drbd_md_put_buffer(mdev);
  2675. out:
  2676. put_ldev(mdev);
  2677. return rv;
  2678. }
  2679. /**
  2680. * drbd_md_mark_dirty() - Mark meta data super block as dirty
  2681. * @mdev: DRBD device.
  2682. *
  2683. * Call this function if you change anything that should be written to
  2684. * the meta-data super block. This function sets MD_DIRTY, and starts a
  2685. * timer that ensures that within five seconds you have to call drbd_md_sync().
  2686. */
  2687. #ifdef DEBUG
  2688. void drbd_md_mark_dirty_(struct drbd_conf *mdev, unsigned int line, const char *func)
  2689. {
  2690. if (!test_and_set_bit(MD_DIRTY, &mdev->flags)) {
  2691. mod_timer(&mdev->md_sync_timer, jiffies + HZ);
  2692. mdev->last_md_mark_dirty.line = line;
  2693. mdev->last_md_mark_dirty.func = func;
  2694. }
  2695. }
  2696. #else
  2697. void drbd_md_mark_dirty(struct drbd_conf *mdev)
  2698. {
  2699. if (!test_and_set_bit(MD_DIRTY, &mdev->flags))
  2700. mod_timer(&mdev->md_sync_timer, jiffies + 5*HZ);
  2701. }
  2702. #endif
  2703. static void drbd_uuid_move_history(struct drbd_conf *mdev) __must_hold(local)
  2704. {
  2705. int i;
  2706. for (i = UI_HISTORY_START; i < UI_HISTORY_END; i++)
  2707. mdev->ldev->md.uuid[i+1] = mdev->ldev->md.uuid[i];
  2708. }
  2709. void _drbd_uuid_set(struct drbd_conf *mdev, int idx, u64 val) __must_hold(local)
  2710. {
  2711. if (idx == UI_CURRENT) {
  2712. if (mdev->state.role == R_PRIMARY)
  2713. val |= 1;
  2714. else
  2715. val &= ~((u64)1);
  2716. drbd_set_ed_uuid(mdev, val);
  2717. }
  2718. mdev->ldev->md.uuid[idx] = val;
  2719. drbd_md_mark_dirty(mdev);
  2720. }
  2721. void drbd_uuid_set(struct drbd_conf *mdev, int idx, u64 val) __must_hold(local)
  2722. {
  2723. if (mdev->ldev->md.uuid[idx]) {
  2724. drbd_uuid_move_history(mdev);
  2725. mdev->ldev->md.uuid[UI_HISTORY_START] = mdev->ldev->md.uuid[idx];
  2726. }
  2727. _drbd_uuid_set(mdev, idx, val);
  2728. }
  2729. /**
  2730. * drbd_uuid_new_current() - Creates a new current UUID
  2731. * @mdev: DRBD device.
  2732. *
  2733. * Creates a new current UUID, and rotates the old current UUID into
  2734. * the bitmap slot. Causes an incremental resync upon next connect.
  2735. */
  2736. void drbd_uuid_new_current(struct drbd_conf *mdev) __must_hold(local)
  2737. {
  2738. u64 val;
  2739. unsigned long long bm_uuid = mdev->ldev->md.uuid[UI_BITMAP];
  2740. if (bm_uuid)
  2741. dev_warn(DEV, "bm UUID was already set: %llX\n", bm_uuid);
  2742. mdev->ldev->md.uuid[UI_BITMAP] = mdev->ldev->md.uuid[UI_CURRENT];
  2743. get_random_bytes(&val, sizeof(u64));
  2744. _drbd_uuid_set(mdev, UI_CURRENT, val);
  2745. drbd_print_uuids(mdev, "new current UUID");
  2746. /* get it to stable storage _now_ */
  2747. drbd_md_sync(mdev);
  2748. }
  2749. void drbd_uuid_set_bm(struct drbd_conf *mdev, u64 val) __must_hold(local)
  2750. {
  2751. if (mdev->ldev->md.uuid[UI_BITMAP] == 0 && val == 0)
  2752. return;
  2753. if (val == 0) {
  2754. drbd_uuid_move_history(mdev);
  2755. mdev->ldev->md.uuid[UI_HISTORY_START] = mdev->ldev->md.uuid[UI_BITMAP];
  2756. mdev->ldev->md.uuid[UI_BITMAP] = 0;
  2757. } else {
  2758. unsigned long long bm_uuid = mdev->ldev->md.uuid[UI_BITMAP];
  2759. if (bm_uuid)
  2760. dev_warn(DEV, "bm UUID was already set: %llX\n", bm_uuid);
  2761. mdev->ldev->md.uuid[UI_BITMAP] = val & ~((u64)1);
  2762. }
  2763. drbd_md_mark_dirty(mdev);
  2764. }
  2765. /**
  2766. * drbd_bmio_set_n_write() - io_fn for drbd_queue_bitmap_io() or drbd_bitmap_io()
  2767. * @mdev: DRBD device.
  2768. *
  2769. * Sets all bits in the bitmap and writes the whole bitmap to stable storage.
  2770. */
  2771. int drbd_bmio_set_n_write(struct drbd_conf *mdev)
  2772. {
  2773. int rv = -EIO;
  2774. if (get_ldev_if_state(mdev, D_ATTACHING)) {
  2775. drbd_md_set_flag(mdev, MDF_FULL_SYNC);
  2776. drbd_md_sync(mdev);
  2777. drbd_bm_set_all(mdev);
  2778. rv = drbd_bm_write(mdev);
  2779. if (!rv) {
  2780. drbd_md_clear_flag(mdev, MDF_FULL_SYNC);
  2781. drbd_md_sync(mdev);
  2782. }
  2783. put_ldev(mdev);
  2784. }
  2785. return rv;
  2786. }
  2787. /**
  2788. * drbd_bmio_clear_n_write() - io_fn for drbd_queue_bitmap_io() or drbd_bitmap_io()
  2789. * @mdev: DRBD device.
  2790. *
  2791. * Clears all bits in the bitmap and writes the whole bitmap to stable storage.
  2792. */
  2793. int drbd_bmio_clear_n_write(struct drbd_conf *mdev)
  2794. {
  2795. int rv = -EIO;
  2796. drbd_resume_al(mdev);
  2797. if (get_ldev_if_state(mdev, D_ATTACHING)) {
  2798. drbd_bm_clear_all(mdev);
  2799. rv = drbd_bm_write(mdev);
  2800. put_ldev(mdev);
  2801. }
  2802. return rv;
  2803. }
  2804. static int w_bitmap_io(struct drbd_work *w, int unused)
  2805. {
  2806. struct bm_io_work *work = container_of(w, struct bm_io_work, w);
  2807. struct drbd_conf *mdev = w->mdev;
  2808. int rv = -EIO;
  2809. D_ASSERT(atomic_read(&mdev->ap_bio_cnt) == 0);
  2810. if (get_ldev(mdev)) {
  2811. drbd_bm_lock(mdev, work->why, work->flags);
  2812. rv = work->io_fn(mdev);
  2813. drbd_bm_unlock(mdev);
  2814. put_ldev(mdev);
  2815. }
  2816. clear_bit_unlock(BITMAP_IO, &mdev->flags);
  2817. wake_up(&mdev->misc_wait);
  2818. if (work->done)
  2819. work->done(mdev, rv);
  2820. clear_bit(BITMAP_IO_QUEUED, &mdev->flags);
  2821. work->why = NULL;
  2822. work->flags = 0;
  2823. return 0;
  2824. }
  2825. void drbd_ldev_destroy(struct drbd_conf *mdev)
  2826. {
  2827. lc_destroy(mdev->resync);
  2828. mdev->resync = NULL;
  2829. lc_destroy(mdev->act_log);
  2830. mdev->act_log = NULL;
  2831. __no_warn(local,
  2832. drbd_free_bc(mdev->ldev);
  2833. mdev->ldev = NULL;);
  2834. clear_bit(GO_DISKLESS, &mdev->flags);
  2835. }
  2836. static int w_go_diskless(struct drbd_work *w, int unused)
  2837. {
  2838. struct drbd_conf *mdev = w->mdev;
  2839. D_ASSERT(mdev->state.disk == D_FAILED);
  2840. /* we cannot assert local_cnt == 0 here, as get_ldev_if_state will
  2841. * inc/dec it frequently. Once we are D_DISKLESS, no one will touch
  2842. * the protected members anymore, though, so once put_ldev reaches zero
  2843. * again, it will be safe to free them. */
  2844. drbd_force_state(mdev, NS(disk, D_DISKLESS));
  2845. return 0;
  2846. }
  2847. void drbd_go_diskless(struct drbd_conf *mdev)
  2848. {
  2849. D_ASSERT(mdev->state.disk == D_FAILED);
  2850. if (!test_and_set_bit(GO_DISKLESS, &mdev->flags))
  2851. drbd_queue_work(&mdev->tconn->data.work, &mdev->go_diskless);
  2852. }
  2853. /**
  2854. * drbd_queue_bitmap_io() - Queues an IO operation on the whole bitmap
  2855. * @mdev: DRBD device.
  2856. * @io_fn: IO callback to be called when bitmap IO is possible
  2857. * @done: callback to be called after the bitmap IO was performed
  2858. * @why: Descriptive text of the reason for doing the IO
  2859. *
  2860. * While IO on the bitmap happens we freeze application IO thus we ensure
  2861. * that drbd_set_out_of_sync() can not be called. This function MAY ONLY be
  2862. * called from worker context. It MUST NOT be used while a previous such
  2863. * work is still pending!
  2864. */
  2865. void drbd_queue_bitmap_io(struct drbd_conf *mdev,
  2866. int (*io_fn)(struct drbd_conf *),
  2867. void (*done)(struct drbd_conf *, int),
  2868. char *why, enum bm_flag flags)
  2869. {
  2870. D_ASSERT(current == mdev->tconn->worker.task);
  2871. D_ASSERT(!test_bit(BITMAP_IO_QUEUED, &mdev->flags));
  2872. D_ASSERT(!test_bit(BITMAP_IO, &mdev->flags));
  2873. D_ASSERT(list_empty(&mdev->bm_io_work.w.list));
  2874. if (mdev->bm_io_work.why)
  2875. dev_err(DEV, "FIXME going to queue '%s' but '%s' still pending?\n",
  2876. why, mdev->bm_io_work.why);
  2877. mdev->bm_io_work.io_fn = io_fn;
  2878. mdev->bm_io_work.done = done;
  2879. mdev->bm_io_work.why = why;
  2880. mdev->bm_io_work.flags = flags;
  2881. spin_lock_irq(&mdev->tconn->req_lock);
  2882. set_bit(BITMAP_IO, &mdev->flags);
  2883. if (atomic_read(&mdev->ap_bio_cnt) == 0) {
  2884. if (!test_and_set_bit(BITMAP_IO_QUEUED, &mdev->flags))
  2885. drbd_queue_work(&mdev->tconn->data.work, &mdev->bm_io_work.w);
  2886. }
  2887. spin_unlock_irq(&mdev->tconn->req_lock);
  2888. }
  2889. /**
  2890. * drbd_bitmap_io() - Does an IO operation on the whole bitmap
  2891. * @mdev: DRBD device.
  2892. * @io_fn: IO callback to be called when bitmap IO is possible
  2893. * @why: Descriptive text of the reason for doing the IO
  2894. *
  2895. * freezes application IO while that the actual IO operations runs. This
  2896. * functions MAY NOT be called from worker context.
  2897. */
  2898. int drbd_bitmap_io(struct drbd_conf *mdev, int (*io_fn)(struct drbd_conf *),
  2899. char *why, enum bm_flag flags)
  2900. {
  2901. int rv;
  2902. D_ASSERT(current != mdev->tconn->worker.task);
  2903. if ((flags & BM_LOCKED_SET_ALLOWED) == 0)
  2904. drbd_suspend_io(mdev);
  2905. drbd_bm_lock(mdev, why, flags);
  2906. rv = io_fn(mdev);
  2907. drbd_bm_unlock(mdev);
  2908. if ((flags & BM_LOCKED_SET_ALLOWED) == 0)
  2909. drbd_resume_io(mdev);
  2910. return rv;
  2911. }
  2912. void drbd_md_set_flag(struct drbd_conf *mdev, int flag) __must_hold(local)
  2913. {
  2914. if ((mdev->ldev->md.flags & flag) != flag) {
  2915. drbd_md_mark_dirty(mdev);
  2916. mdev->ldev->md.flags |= flag;
  2917. }
  2918. }
  2919. void drbd_md_clear_flag(struct drbd_conf *mdev, int flag) __must_hold(local)
  2920. {
  2921. if ((mdev->ldev->md.flags & flag) != 0) {
  2922. drbd_md_mark_dirty(mdev);
  2923. mdev->ldev->md.flags &= ~flag;
  2924. }
  2925. }
  2926. int drbd_md_test_flag(struct drbd_backing_dev *bdev, int flag)
  2927. {
  2928. return (bdev->md.flags & flag) != 0;
  2929. }
  2930. static void md_sync_timer_fn(unsigned long data)
  2931. {
  2932. struct drbd_conf *mdev = (struct drbd_conf *) data;
  2933. drbd_queue_work_front(&mdev->tconn->data.work, &mdev->md_sync_work);
  2934. }
  2935. static int w_md_sync(struct drbd_work *w, int unused)
  2936. {
  2937. struct drbd_conf *mdev = w->mdev;
  2938. dev_warn(DEV, "md_sync_timer expired! Worker calls drbd_md_sync().\n");
  2939. #ifdef DEBUG
  2940. dev_warn(DEV, "last md_mark_dirty: %s:%u\n",
  2941. mdev->last_md_mark_dirty.func, mdev->last_md_mark_dirty.line);
  2942. #endif
  2943. drbd_md_sync(mdev);
  2944. return 0;
  2945. }
  2946. const char *cmdname(enum drbd_packet cmd)
  2947. {
  2948. /* THINK may need to become several global tables
  2949. * when we want to support more than
  2950. * one PRO_VERSION */
  2951. static const char *cmdnames[] = {
  2952. [P_DATA] = "Data",
  2953. [P_DATA_REPLY] = "DataReply",
  2954. [P_RS_DATA_REPLY] = "RSDataReply",
  2955. [P_BARRIER] = "Barrier",
  2956. [P_BITMAP] = "ReportBitMap",
  2957. [P_BECOME_SYNC_TARGET] = "BecomeSyncTarget",
  2958. [P_BECOME_SYNC_SOURCE] = "BecomeSyncSource",
  2959. [P_UNPLUG_REMOTE] = "UnplugRemote",
  2960. [P_DATA_REQUEST] = "DataRequest",
  2961. [P_RS_DATA_REQUEST] = "RSDataRequest",
  2962. [P_SYNC_PARAM] = "SyncParam",
  2963. [P_SYNC_PARAM89] = "SyncParam89",
  2964. [P_PROTOCOL] = "ReportProtocol",
  2965. [P_UUIDS] = "ReportUUIDs",
  2966. [P_SIZES] = "ReportSizes",
  2967. [P_STATE] = "ReportState",
  2968. [P_SYNC_UUID] = "ReportSyncUUID",
  2969. [P_AUTH_CHALLENGE] = "AuthChallenge",
  2970. [P_AUTH_RESPONSE] = "AuthResponse",
  2971. [P_PING] = "Ping",
  2972. [P_PING_ACK] = "PingAck",
  2973. [P_RECV_ACK] = "RecvAck",
  2974. [P_WRITE_ACK] = "WriteAck",
  2975. [P_RS_WRITE_ACK] = "RSWriteAck",
  2976. [P_DISCARD_WRITE] = "DiscardWrite",
  2977. [P_NEG_ACK] = "NegAck",
  2978. [P_NEG_DREPLY] = "NegDReply",
  2979. [P_NEG_RS_DREPLY] = "NegRSDReply",
  2980. [P_BARRIER_ACK] = "BarrierAck",
  2981. [P_STATE_CHG_REQ] = "StateChgRequest",
  2982. [P_STATE_CHG_REPLY] = "StateChgReply",
  2983. [P_OV_REQUEST] = "OVRequest",
  2984. [P_OV_REPLY] = "OVReply",
  2985. [P_OV_RESULT] = "OVResult",
  2986. [P_CSUM_RS_REQUEST] = "CsumRSRequest",
  2987. [P_RS_IS_IN_SYNC] = "CsumRSIsInSync",
  2988. [P_COMPRESSED_BITMAP] = "CBitmap",
  2989. [P_DELAY_PROBE] = "DelayProbe",
  2990. [P_OUT_OF_SYNC] = "OutOfSync",
  2991. [P_RETRY_WRITE] = "RetryWrite",
  2992. [P_RS_CANCEL] = "RSCancel",
  2993. [P_CONN_ST_CHG_REQ] = "conn_st_chg_req",
  2994. [P_CONN_ST_CHG_REPLY] = "conn_st_chg_reply",
  2995. [P_RETRY_WRITE] = "retry_write",
  2996. [P_PROTOCOL_UPDATE] = "protocol_update",
  2997. /* enum drbd_packet, but not commands - obsoleted flags:
  2998. * P_MAY_IGNORE
  2999. * P_MAX_OPT_CMD
  3000. */
  3001. };
  3002. /* too big for the array: 0xfffX */
  3003. if (cmd == P_INITIAL_META)
  3004. return "InitialMeta";
  3005. if (cmd == P_INITIAL_DATA)
  3006. return "InitialData";
  3007. if (cmd == P_CONNECTION_FEATURES)
  3008. return "ConnectionFeatures";
  3009. if (cmd >= ARRAY_SIZE(cmdnames))
  3010. return "Unknown";
  3011. return cmdnames[cmd];
  3012. }
  3013. /**
  3014. * drbd_wait_misc - wait for a request to make progress
  3015. * @mdev: device associated with the request
  3016. * @i: the struct drbd_interval embedded in struct drbd_request or
  3017. * struct drbd_peer_request
  3018. */
  3019. int drbd_wait_misc(struct drbd_conf *mdev, struct drbd_interval *i)
  3020. {
  3021. struct net_conf *nc;
  3022. DEFINE_WAIT(wait);
  3023. long timeout;
  3024. rcu_read_lock();
  3025. nc = rcu_dereference(mdev->tconn->net_conf);
  3026. if (!nc) {
  3027. rcu_read_unlock();
  3028. return -ETIMEDOUT;
  3029. }
  3030. timeout = nc->ko_count ? nc->timeout * HZ / 10 * nc->ko_count : MAX_SCHEDULE_TIMEOUT;
  3031. rcu_read_unlock();
  3032. /* Indicate to wake up mdev->misc_wait on progress. */
  3033. i->waiting = true;
  3034. prepare_to_wait(&mdev->misc_wait, &wait, TASK_INTERRUPTIBLE);
  3035. spin_unlock_irq(&mdev->tconn->req_lock);
  3036. timeout = schedule_timeout(timeout);
  3037. finish_wait(&mdev->misc_wait, &wait);
  3038. spin_lock_irq(&mdev->tconn->req_lock);
  3039. if (!timeout || mdev->state.conn < C_CONNECTED)
  3040. return -ETIMEDOUT;
  3041. if (signal_pending(current))
  3042. return -ERESTARTSYS;
  3043. return 0;
  3044. }
  3045. #ifdef CONFIG_DRBD_FAULT_INJECTION
  3046. /* Fault insertion support including random number generator shamelessly
  3047. * stolen from kernel/rcutorture.c */
  3048. struct fault_random_state {
  3049. unsigned long state;
  3050. unsigned long count;
  3051. };
  3052. #define FAULT_RANDOM_MULT 39916801 /* prime */
  3053. #define FAULT_RANDOM_ADD 479001701 /* prime */
  3054. #define FAULT_RANDOM_REFRESH 10000
  3055. /*
  3056. * Crude but fast random-number generator. Uses a linear congruential
  3057. * generator, with occasional help from get_random_bytes().
  3058. */
  3059. static unsigned long
  3060. _drbd_fault_random(struct fault_random_state *rsp)
  3061. {
  3062. long refresh;
  3063. if (!rsp->count--) {
  3064. get_random_bytes(&refresh, sizeof(refresh));
  3065. rsp->state += refresh;
  3066. rsp->count = FAULT_RANDOM_REFRESH;
  3067. }
  3068. rsp->state = rsp->state * FAULT_RANDOM_MULT + FAULT_RANDOM_ADD;
  3069. return swahw32(rsp->state);
  3070. }
  3071. static char *
  3072. _drbd_fault_str(unsigned int type) {
  3073. static char *_faults[] = {
  3074. [DRBD_FAULT_MD_WR] = "Meta-data write",
  3075. [DRBD_FAULT_MD_RD] = "Meta-data read",
  3076. [DRBD_FAULT_RS_WR] = "Resync write",
  3077. [DRBD_FAULT_RS_RD] = "Resync read",
  3078. [DRBD_FAULT_DT_WR] = "Data write",
  3079. [DRBD_FAULT_DT_RD] = "Data read",
  3080. [DRBD_FAULT_DT_RA] = "Data read ahead",
  3081. [DRBD_FAULT_BM_ALLOC] = "BM allocation",
  3082. [DRBD_FAULT_AL_EE] = "EE allocation",
  3083. [DRBD_FAULT_RECEIVE] = "receive data corruption",
  3084. };
  3085. return (type < DRBD_FAULT_MAX) ? _faults[type] : "**Unknown**";
  3086. }
  3087. unsigned int
  3088. _drbd_insert_fault(struct drbd_conf *mdev, unsigned int type)
  3089. {
  3090. static struct fault_random_state rrs = {0, 0};
  3091. unsigned int ret = (
  3092. (fault_devs == 0 ||
  3093. ((1 << mdev_to_minor(mdev)) & fault_devs) != 0) &&
  3094. (((_drbd_fault_random(&rrs) % 100) + 1) <= fault_rate));
  3095. if (ret) {
  3096. fault_count++;
  3097. if (__ratelimit(&drbd_ratelimit_state))
  3098. dev_warn(DEV, "***Simulating %s failure\n",
  3099. _drbd_fault_str(type));
  3100. }
  3101. return ret;
  3102. }
  3103. #endif
  3104. const char *drbd_buildtag(void)
  3105. {
  3106. /* DRBD built from external sources has here a reference to the
  3107. git hash of the source code. */
  3108. static char buildtag[38] = "\0uilt-in";
  3109. if (buildtag[0] == 0) {
  3110. #ifdef CONFIG_MODULES
  3111. if (THIS_MODULE != NULL)
  3112. sprintf(buildtag, "srcversion: %-24s", THIS_MODULE->srcversion);
  3113. else
  3114. #endif
  3115. buildtag[0] = 'b';
  3116. }
  3117. return buildtag;
  3118. }
  3119. module_init(drbd_init)
  3120. module_exit(drbd_cleanup)
  3121. EXPORT_SYMBOL(drbd_conn_str);
  3122. EXPORT_SYMBOL(drbd_role_str);
  3123. EXPORT_SYMBOL(drbd_disk_str);
  3124. EXPORT_SYMBOL(drbd_set_st_err_str);