@@ -212,7 +212,8 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
212212 dst_page -- ;
213213 }
214214 src -= len ;
215- memmove (kmap (* dst_page ) + src , kmap (* src_page ) + src , len );
215+ memmove (kmap (* dst_page ) + src ,
216+ kmap (* src_page ) + src , len );
216217 kunmap (* src_page );
217218 set_page_dirty (* dst_page );
218219 kunmap (* dst_page );
@@ -250,14 +251,16 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
250251
251252 if (src == dst ) {
252253 l = min (len , (int )PAGE_CACHE_SIZE - src );
253- memmove (kmap (* dst_page ) + src , kmap (* src_page ) + src , l );
254+ memmove (kmap (* dst_page ) + src ,
255+ kmap (* src_page ) + src , l );
254256 kunmap (* src_page );
255257 set_page_dirty (* dst_page );
256258 kunmap (* dst_page );
257259
258260 while ((len -= l ) != 0 ) {
259261 l = min (len , (int )PAGE_CACHE_SIZE );
260- memmove (kmap (* ++ dst_page ), kmap (* ++ src_page ), l );
262+ memmove (kmap (* ++ dst_page ),
263+ kmap (* ++ src_page ), l );
261264 kunmap (* src_page );
262265 set_page_dirty (* dst_page );
263266 kunmap (* dst_page );
@@ -268,7 +271,8 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
268271 do {
269272 src_ptr = kmap (* src_page ) + src ;
270273 dst_ptr = kmap (* dst_page ) + dst ;
271- if (PAGE_CACHE_SIZE - src < PAGE_CACHE_SIZE - dst ) {
274+ if (PAGE_CACHE_SIZE - src <
275+ PAGE_CACHE_SIZE - dst ) {
272276 l = PAGE_CACHE_SIZE - src ;
273277 src = 0 ;
274278 dst += l ;
@@ -340,7 +344,8 @@ void hfs_bnode_unlink(struct hfs_bnode *node)
340344 return ;
341345 tmp -> next = node -> next ;
342346 cnid = cpu_to_be32 (tmp -> next );
343- hfs_bnode_write (tmp , & cnid , offsetof(struct hfs_bnode_desc , next ), 4 );
347+ hfs_bnode_write (tmp , & cnid ,
348+ offsetof(struct hfs_bnode_desc , next ), 4 );
344349 hfs_bnode_put (tmp );
345350 } else if (node -> type == HFS_NODE_LEAF )
346351 tree -> leaf_head = node -> next ;
@@ -351,7 +356,8 @@ void hfs_bnode_unlink(struct hfs_bnode *node)
351356 return ;
352357 tmp -> prev = node -> prev ;
353358 cnid = cpu_to_be32 (tmp -> prev );
354- hfs_bnode_write (tmp , & cnid , offsetof(struct hfs_bnode_desc , prev ), 4 );
359+ hfs_bnode_write (tmp , & cnid ,
360+ offsetof(struct hfs_bnode_desc , prev ), 4 );
355361 hfs_bnode_put (tmp );
356362 } else if (node -> type == HFS_NODE_LEAF )
357363 tree -> leaf_tail = node -> prev ;
@@ -379,7 +385,9 @@ struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid)
379385 struct hfs_bnode * node ;
380386
381387 if (cnid >= tree -> node_count ) {
382- printk (KERN_ERR "hfs: request for non-existent node %d in B*Tree\n" , cnid );
388+ printk (KERN_ERR "hfs: request for non-existent node "
389+ "%d in B*Tree\n" ,
390+ cnid );
383391 return NULL ;
384392 }
385393
@@ -402,7 +410,9 @@ static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
402410 loff_t off ;
403411
404412 if (cnid >= tree -> node_count ) {
405- printk (KERN_ERR "hfs: request for non-existent node %d in B*Tree\n" , cnid );
413+ printk (KERN_ERR "hfs: request for non-existent node "
414+ "%d in B*Tree\n" ,
415+ cnid );
406416 return NULL ;
407417 }
408418
@@ -429,7 +439,8 @@ static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
429439 } else {
430440 spin_unlock (& tree -> hash_lock );
431441 kfree (node );
432- wait_event (node2 -> lock_wq , !test_bit (HFS_BNODE_NEW , & node2 -> flags ));
442+ wait_event (node2 -> lock_wq ,
443+ !test_bit (HFS_BNODE_NEW , & node2 -> flags ));
433444 return node2 ;
434445 }
435446 spin_unlock (& tree -> hash_lock );
@@ -483,7 +494,8 @@ struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
483494 if (node ) {
484495 hfs_bnode_get (node );
485496 spin_unlock (& tree -> hash_lock );
486- wait_event (node -> lock_wq , !test_bit (HFS_BNODE_NEW , & node -> flags ));
497+ wait_event (node -> lock_wq ,
498+ !test_bit (HFS_BNODE_NEW , & node -> flags ));
487499 if (test_bit (HFS_BNODE_ERROR , & node -> flags ))
488500 goto node_error ;
489501 return node ;
@@ -497,7 +509,8 @@ struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
497509 if (!test_bit (HFS_BNODE_NEW , & node -> flags ))
498510 return node ;
499511
500- desc = (struct hfs_bnode_desc * )(kmap (node -> page [0 ]) + node -> page_offset );
512+ desc = (struct hfs_bnode_desc * )(kmap (node -> page [0 ]) +
513+ node -> page_offset );
501514 node -> prev = be32_to_cpu (desc -> prev );
502515 node -> next = be32_to_cpu (desc -> next );
503516 node -> num_recs = be16_to_cpu (desc -> num_recs );
@@ -607,7 +620,8 @@ void hfs_bnode_get(struct hfs_bnode *node)
607620 if (node ) {
608621 atomic_inc (& node -> refcnt );
609622 dprint (DBG_BNODE_REFS , "get_node(%d:%d): %d\n" ,
610- node -> tree -> cnid , node -> this , atomic_read (& node -> refcnt ));
623+ node -> tree -> cnid , node -> this ,
624+ atomic_read (& node -> refcnt ));
611625 }
612626}
613627
@@ -619,7 +633,8 @@ void hfs_bnode_put(struct hfs_bnode *node)
619633 int i ;
620634
621635 dprint (DBG_BNODE_REFS , "put_node(%d:%d): %d\n" ,
622- node -> tree -> cnid , node -> this , atomic_read (& node -> refcnt ));
636+ node -> tree -> cnid , node -> this ,
637+ atomic_read (& node -> refcnt ));
623638 BUG_ON (!atomic_read (& node -> refcnt ));
624639 if (!atomic_dec_and_lock (& node -> refcnt , & tree -> hash_lock ))
625640 return ;
0 commit comments