diff options
author | Harvey Harrison <harvey.harrison@gmail.com> | 2008-04-28 05:15:50 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-04-28 11:58:42 -0400 |
commit | e46b272b6608783ed7aa7b0594871550ce20b849 (patch) | |
tree | a65834f60bee7e77f9669a7ff7e95a1ec3cd74ee | |
parent | 9a7b2b0f36c40beeb252cb5ceff36bb295e88d97 (diff) |
md: replace remaining __FUNCTION__ occurrences
__FUNCTION__ is gcc-specific, use __func__
Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com>
Cc: Neil Brown <neilb@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | drivers/md/dm-uevent.c | 22 | ||||
-rw-r--r-- | drivers/md/raid5.c | 28 |
2 files changed, 25 insertions, 25 deletions
diff --git a/drivers/md/dm-uevent.c b/drivers/md/dm-uevent.c index 50377e5dc2a3..6f65883aef12 100644 --- a/drivers/md/dm-uevent.c +++ b/drivers/md/dm-uevent.c | |||
@@ -78,7 +78,7 @@ static struct dm_uevent *dm_build_path_uevent(struct mapped_device *md, | |||
78 | 78 | ||
79 | event = dm_uevent_alloc(md); | 79 | event = dm_uevent_alloc(md); |
80 | if (!event) { | 80 | if (!event) { |
81 | DMERR("%s: dm_uevent_alloc() failed", __FUNCTION__); | 81 | DMERR("%s: dm_uevent_alloc() failed", __func__); |
82 | goto err_nomem; | 82 | goto err_nomem; |
83 | } | 83 | } |
84 | 84 | ||
@@ -86,32 +86,32 @@ static struct dm_uevent *dm_build_path_uevent(struct mapped_device *md, | |||
86 | 86 | ||
87 | if (add_uevent_var(&event->ku_env, "DM_TARGET=%s", ti->type->name)) { | 87 | if (add_uevent_var(&event->ku_env, "DM_TARGET=%s", ti->type->name)) { |
88 | DMERR("%s: add_uevent_var() for DM_TARGET failed", | 88 | DMERR("%s: add_uevent_var() for DM_TARGET failed", |
89 | __FUNCTION__); | 89 | __func__); |
90 | goto err_add; | 90 | goto err_add; |
91 | } | 91 | } |
92 | 92 | ||
93 | if (add_uevent_var(&event->ku_env, "DM_ACTION=%s", dm_action)) { | 93 | if (add_uevent_var(&event->ku_env, "DM_ACTION=%s", dm_action)) { |
94 | DMERR("%s: add_uevent_var() for DM_ACTION failed", | 94 | DMERR("%s: add_uevent_var() for DM_ACTION failed", |
95 | __FUNCTION__); | 95 | __func__); |
96 | goto err_add; | 96 | goto err_add; |
97 | } | 97 | } |
98 | 98 | ||
99 | if (add_uevent_var(&event->ku_env, "DM_SEQNUM=%u", | 99 | if (add_uevent_var(&event->ku_env, "DM_SEQNUM=%u", |
100 | dm_next_uevent_seq(md))) { | 100 | dm_next_uevent_seq(md))) { |
101 | DMERR("%s: add_uevent_var() for DM_SEQNUM failed", | 101 | DMERR("%s: add_uevent_var() for DM_SEQNUM failed", |
102 | __FUNCTION__); | 102 | __func__); |
103 | goto err_add; | 103 | goto err_add; |
104 | } | 104 | } |
105 | 105 | ||
106 | if (add_uevent_var(&event->ku_env, "DM_PATH=%s", path)) { | 106 | if (add_uevent_var(&event->ku_env, "DM_PATH=%s", path)) { |
107 | DMERR("%s: add_uevent_var() for DM_PATH failed", __FUNCTION__); | 107 | DMERR("%s: add_uevent_var() for DM_PATH failed", __func__); |
108 | goto err_add; | 108 | goto err_add; |
109 | } | 109 | } |
110 | 110 | ||
111 | if (add_uevent_var(&event->ku_env, "DM_NR_VALID_PATHS=%d", | 111 | if (add_uevent_var(&event->ku_env, "DM_NR_VALID_PATHS=%d", |
112 | nr_valid_paths)) { | 112 | nr_valid_paths)) { |
113 | DMERR("%s: add_uevent_var() for DM_NR_VALID_PATHS failed", | 113 | DMERR("%s: add_uevent_var() for DM_NR_VALID_PATHS failed", |
114 | __FUNCTION__); | 114 | __func__); |
115 | goto err_add; | 115 | goto err_add; |
116 | } | 116 | } |
117 | 117 | ||
@@ -146,25 +146,25 @@ void dm_send_uevents(struct list_head *events, struct kobject *kobj) | |||
146 | if (dm_copy_name_and_uuid(event->md, event->name, | 146 | if (dm_copy_name_and_uuid(event->md, event->name, |
147 | event->uuid)) { | 147 | event->uuid)) { |
148 | DMERR("%s: dm_copy_name_and_uuid() failed", | 148 | DMERR("%s: dm_copy_name_and_uuid() failed", |
149 | __FUNCTION__); | 149 | __func__); |
150 | goto uevent_free; | 150 | goto uevent_free; |
151 | } | 151 | } |
152 | 152 | ||
153 | if (add_uevent_var(&event->ku_env, "DM_NAME=%s", event->name)) { | 153 | if (add_uevent_var(&event->ku_env, "DM_NAME=%s", event->name)) { |
154 | DMERR("%s: add_uevent_var() for DM_NAME failed", | 154 | DMERR("%s: add_uevent_var() for DM_NAME failed", |
155 | __FUNCTION__); | 155 | __func__); |
156 | goto uevent_free; | 156 | goto uevent_free; |
157 | } | 157 | } |
158 | 158 | ||
159 | if (add_uevent_var(&event->ku_env, "DM_UUID=%s", event->uuid)) { | 159 | if (add_uevent_var(&event->ku_env, "DM_UUID=%s", event->uuid)) { |
160 | DMERR("%s: add_uevent_var() for DM_UUID failed", | 160 | DMERR("%s: add_uevent_var() for DM_UUID failed", |
161 | __FUNCTION__); | 161 | __func__); |
162 | goto uevent_free; | 162 | goto uevent_free; |
163 | } | 163 | } |
164 | 164 | ||
165 | r = kobject_uevent_env(kobj, event->action, event->ku_env.envp); | 165 | r = kobject_uevent_env(kobj, event->action, event->ku_env.envp); |
166 | if (r) | 166 | if (r) |
167 | DMERR("%s: kobject_uevent_env failed", __FUNCTION__); | 167 | DMERR("%s: kobject_uevent_env failed", __func__); |
168 | uevent_free: | 168 | uevent_free: |
169 | dm_uevent_free(event); | 169 | dm_uevent_free(event); |
170 | } | 170 | } |
@@ -187,7 +187,7 @@ void dm_path_uevent(enum dm_uevent_type event_type, struct dm_target *ti, | |||
187 | struct dm_uevent *event; | 187 | struct dm_uevent *event; |
188 | 188 | ||
189 | if (event_type >= ARRAY_SIZE(_dm_uevent_type_names)) { | 189 | if (event_type >= ARRAY_SIZE(_dm_uevent_type_names)) { |
190 | DMERR("%s: Invalid event_type %d", __FUNCTION__, event_type); | 190 | DMERR("%s: Invalid event_type %d", __func__, event_type); |
191 | goto out; | 191 | goto out; |
192 | } | 192 | } |
193 | 193 | ||
diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c index b162b839a662..4efec467e2f1 100644 --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c | |||
@@ -433,7 +433,7 @@ static void ops_run_io(struct stripe_head *sh) | |||
433 | 433 | ||
434 | bi->bi_bdev = rdev->bdev; | 434 | bi->bi_bdev = rdev->bdev; |
435 | pr_debug("%s: for %llu schedule op %ld on disc %d\n", | 435 | pr_debug("%s: for %llu schedule op %ld on disc %d\n", |
436 | __FUNCTION__, (unsigned long long)sh->sector, | 436 | __func__, (unsigned long long)sh->sector, |
437 | bi->bi_rw, i); | 437 | bi->bi_rw, i); |
438 | atomic_inc(&sh->count); | 438 | atomic_inc(&sh->count); |
439 | bi->bi_sector = sh->sector + rdev->data_offset; | 439 | bi->bi_sector = sh->sector + rdev->data_offset; |
@@ -520,7 +520,7 @@ static void ops_complete_biofill(void *stripe_head_ref) | |||
520 | raid5_conf_t *conf = sh->raid_conf; | 520 | raid5_conf_t *conf = sh->raid_conf; |
521 | int i; | 521 | int i; |
522 | 522 | ||
523 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 523 | pr_debug("%s: stripe %llu\n", __func__, |
524 | (unsigned long long)sh->sector); | 524 | (unsigned long long)sh->sector); |
525 | 525 | ||
526 | /* clear completed biofills */ | 526 | /* clear completed biofills */ |
@@ -569,7 +569,7 @@ static void ops_run_biofill(struct stripe_head *sh) | |||
569 | raid5_conf_t *conf = sh->raid_conf; | 569 | raid5_conf_t *conf = sh->raid_conf; |
570 | int i; | 570 | int i; |
571 | 571 | ||
572 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 572 | pr_debug("%s: stripe %llu\n", __func__, |
573 | (unsigned long long)sh->sector); | 573 | (unsigned long long)sh->sector); |
574 | 574 | ||
575 | for (i = sh->disks; i--; ) { | 575 | for (i = sh->disks; i--; ) { |
@@ -600,7 +600,7 @@ static void ops_complete_compute5(void *stripe_head_ref) | |||
600 | int target = sh->ops.target; | 600 | int target = sh->ops.target; |
601 | struct r5dev *tgt = &sh->dev[target]; | 601 | struct r5dev *tgt = &sh->dev[target]; |
602 | 602 | ||
603 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 603 | pr_debug("%s: stripe %llu\n", __func__, |
604 | (unsigned long long)sh->sector); | 604 | (unsigned long long)sh->sector); |
605 | 605 | ||
606 | set_bit(R5_UPTODATE, &tgt->flags); | 606 | set_bit(R5_UPTODATE, &tgt->flags); |
@@ -625,7 +625,7 @@ ops_run_compute5(struct stripe_head *sh, unsigned long pending) | |||
625 | int i; | 625 | int i; |
626 | 626 | ||
627 | pr_debug("%s: stripe %llu block: %d\n", | 627 | pr_debug("%s: stripe %llu block: %d\n", |
628 | __FUNCTION__, (unsigned long long)sh->sector, target); | 628 | __func__, (unsigned long long)sh->sector, target); |
629 | BUG_ON(!test_bit(R5_Wantcompute, &tgt->flags)); | 629 | BUG_ON(!test_bit(R5_Wantcompute, &tgt->flags)); |
630 | 630 | ||
631 | for (i = disks; i--; ) | 631 | for (i = disks; i--; ) |
@@ -653,7 +653,7 @@ static void ops_complete_prexor(void *stripe_head_ref) | |||
653 | { | 653 | { |
654 | struct stripe_head *sh = stripe_head_ref; | 654 | struct stripe_head *sh = stripe_head_ref; |
655 | 655 | ||
656 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 656 | pr_debug("%s: stripe %llu\n", __func__, |
657 | (unsigned long long)sh->sector); | 657 | (unsigned long long)sh->sector); |
658 | 658 | ||
659 | set_bit(STRIPE_OP_PREXOR, &sh->ops.complete); | 659 | set_bit(STRIPE_OP_PREXOR, &sh->ops.complete); |
@@ -670,7 +670,7 @@ ops_run_prexor(struct stripe_head *sh, struct dma_async_tx_descriptor *tx) | |||
670 | /* existing parity data subtracted */ | 670 | /* existing parity data subtracted */ |
671 | struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; | 671 | struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; |
672 | 672 | ||
673 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 673 | pr_debug("%s: stripe %llu\n", __func__, |
674 | (unsigned long long)sh->sector); | 674 | (unsigned long long)sh->sector); |
675 | 675 | ||
676 | for (i = disks; i--; ) { | 676 | for (i = disks; i--; ) { |
@@ -699,7 +699,7 @@ ops_run_biodrain(struct stripe_head *sh, struct dma_async_tx_descriptor *tx, | |||
699 | */ | 699 | */ |
700 | int prexor = test_bit(STRIPE_OP_PREXOR, &pending); | 700 | int prexor = test_bit(STRIPE_OP_PREXOR, &pending); |
701 | 701 | ||
702 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 702 | pr_debug("%s: stripe %llu\n", __func__, |
703 | (unsigned long long)sh->sector); | 703 | (unsigned long long)sh->sector); |
704 | 704 | ||
705 | for (i = disks; i--; ) { | 705 | for (i = disks; i--; ) { |
@@ -744,7 +744,7 @@ static void ops_complete_postxor(void *stripe_head_ref) | |||
744 | { | 744 | { |
745 | struct stripe_head *sh = stripe_head_ref; | 745 | struct stripe_head *sh = stripe_head_ref; |
746 | 746 | ||
747 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 747 | pr_debug("%s: stripe %llu\n", __func__, |
748 | (unsigned long long)sh->sector); | 748 | (unsigned long long)sh->sector); |
749 | 749 | ||
750 | set_bit(STRIPE_OP_POSTXOR, &sh->ops.complete); | 750 | set_bit(STRIPE_OP_POSTXOR, &sh->ops.complete); |
@@ -757,7 +757,7 @@ static void ops_complete_write(void *stripe_head_ref) | |||
757 | struct stripe_head *sh = stripe_head_ref; | 757 | struct stripe_head *sh = stripe_head_ref; |
758 | int disks = sh->disks, i, pd_idx = sh->pd_idx; | 758 | int disks = sh->disks, i, pd_idx = sh->pd_idx; |
759 | 759 | ||
760 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 760 | pr_debug("%s: stripe %llu\n", __func__, |
761 | (unsigned long long)sh->sector); | 761 | (unsigned long long)sh->sector); |
762 | 762 | ||
763 | for (i = disks; i--; ) { | 763 | for (i = disks; i--; ) { |
@@ -787,7 +787,7 @@ ops_run_postxor(struct stripe_head *sh, struct dma_async_tx_descriptor *tx, | |||
787 | unsigned long flags; | 787 | unsigned long flags; |
788 | dma_async_tx_callback callback; | 788 | dma_async_tx_callback callback; |
789 | 789 | ||
790 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 790 | pr_debug("%s: stripe %llu\n", __func__, |
791 | (unsigned long long)sh->sector); | 791 | (unsigned long long)sh->sector); |
792 | 792 | ||
793 | /* check if prexor is active which means only process blocks | 793 | /* check if prexor is active which means only process blocks |
@@ -837,7 +837,7 @@ static void ops_complete_check(void *stripe_head_ref) | |||
837 | struct stripe_head *sh = stripe_head_ref; | 837 | struct stripe_head *sh = stripe_head_ref; |
838 | int pd_idx = sh->pd_idx; | 838 | int pd_idx = sh->pd_idx; |
839 | 839 | ||
840 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 840 | pr_debug("%s: stripe %llu\n", __func__, |
841 | (unsigned long long)sh->sector); | 841 | (unsigned long long)sh->sector); |
842 | 842 | ||
843 | if (test_and_clear_bit(STRIPE_OP_MOD_DMA_CHECK, &sh->ops.pending) && | 843 | if (test_and_clear_bit(STRIPE_OP_MOD_DMA_CHECK, &sh->ops.pending) && |
@@ -859,7 +859,7 @@ static void ops_run_check(struct stripe_head *sh) | |||
859 | int count = 0, pd_idx = sh->pd_idx, i; | 859 | int count = 0, pd_idx = sh->pd_idx, i; |
860 | struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; | 860 | struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page; |
861 | 861 | ||
862 | pr_debug("%s: stripe %llu\n", __FUNCTION__, | 862 | pr_debug("%s: stripe %llu\n", __func__, |
863 | (unsigned long long)sh->sector); | 863 | (unsigned long long)sh->sector); |
864 | 864 | ||
865 | for (i = disks; i--; ) { | 865 | for (i = disks; i--; ) { |
@@ -1759,7 +1759,7 @@ handle_write_operations5(struct stripe_head *sh, int rcw, int expand) | |||
1759 | locked++; | 1759 | locked++; |
1760 | 1760 | ||
1761 | pr_debug("%s: stripe %llu locked: %d pending: %lx\n", | 1761 | pr_debug("%s: stripe %llu locked: %d pending: %lx\n", |
1762 | __FUNCTION__, (unsigned long long)sh->sector, | 1762 | __func__, (unsigned long long)sh->sector, |
1763 | locked, sh->ops.pending); | 1763 | locked, sh->ops.pending); |
1764 | 1764 | ||
1765 | return locked; | 1765 | return locked; |