add patches for 2.6.23 on brcm47xx (not enabled yet)
[openwrt/svn-archive/archive.git] / target / linux / brcm47xx / patches-2.6.23 / 200-b44_ssb_fixup.patch
1 Index: linux-2.6.23/drivers/net/b44.c
2 ===================================================================
3 --- linux-2.6.23.orig/drivers/net/b44.c 2007-10-13 02:46:38.946989430 +0200
4 +++ linux-2.6.23/drivers/net/b44.c 2007-10-13 03:15:34.889915180 +0200
5 @@ -129,7 +129,7 @@
6 unsigned long offset,
7 enum dma_data_direction dir)
8 {
9 - dma_sync_single_range_for_device(&sdev->dev, dma_base,
10 + dma_sync_single_range_for_device(sdev->dev, dma_base,
11 offset & dma_desc_align_mask,
12 dma_desc_sync_size, dir);
13 }
14 @@ -139,7 +139,7 @@
15 unsigned long offset,
16 enum dma_data_direction dir)
17 {
18 - dma_sync_single_range_for_cpu(&sdev->dev, dma_base,
19 + dma_sync_single_range_for_cpu(sdev->dev, dma_base,
20 offset & dma_desc_align_mask,
21 dma_desc_sync_size, dir);
22 }
23 @@ -563,7 +563,7 @@
24
25 BUG_ON(skb == NULL);
26
27 - dma_unmap_single(&bp->sdev->dev,
28 + dma_unmap_single(bp->sdev->dev,
29 pci_unmap_addr(rp, mapping),
30 skb->len,
31 DMA_TO_DEVICE);
32 @@ -603,7 +603,7 @@
33 if (skb == NULL)
34 return -ENOMEM;
35
36 - mapping = dma_map_single(&bp->sdev->dev, skb->data,
37 + mapping = dma_map_single(bp->sdev->dev, skb->data,
38 RX_PKT_BUF_SZ,
39 DMA_FROM_DEVICE);
40
41 @@ -613,18 +613,18 @@
42 mapping + RX_PKT_BUF_SZ > DMA_30BIT_MASK) {
43 /* Sigh... */
44 if (!dma_mapping_error(mapping))
45 - dma_unmap_single(&bp->sdev->dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE);
46 + dma_unmap_single(bp->sdev->dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE);
47 dev_kfree_skb_any(skb);
48 skb = __netdev_alloc_skb(bp->dev, RX_PKT_BUF_SZ, GFP_ATOMIC|GFP_DMA);
49 if (skb == NULL)
50 return -ENOMEM;
51 - mapping = dma_map_single(&bp->sdev->dev, skb->data,
52 + mapping = dma_map_single(bp->sdev->dev, skb->data,
53 RX_PKT_BUF_SZ,
54 DMA_FROM_DEVICE);
55 if (dma_mapping_error(mapping) ||
56 mapping + RX_PKT_BUF_SZ > DMA_30BIT_MASK) {
57 if (!dma_mapping_error(mapping))
58 - dma_unmap_single(&bp->sdev->dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE);
59 + dma_unmap_single(bp->sdev->dev, mapping, RX_PKT_BUF_SZ,DMA_FROM_DEVICE);
60 dev_kfree_skb_any(skb);
61 return -ENOMEM;
62 }
63 @@ -700,7 +700,7 @@
64 dest_idx * sizeof(dest_desc),
65 DMA_BIDIRECTIONAL);
66
67 - dma_sync_single_for_device(&bp->sdev->dev, le32_to_cpu(src_desc->addr),
68 + dma_sync_single_for_device(bp->sdev->dev, le32_to_cpu(src_desc->addr),
69 RX_PKT_BUF_SZ,
70 DMA_FROM_DEVICE);
71 }
72 @@ -722,7 +722,7 @@
73 struct rx_header *rh;
74 u16 len;
75
76 - dma_sync_single_for_cpu(&bp->sdev->dev, map,
77 + dma_sync_single_for_cpu(bp->sdev->dev, map,
78 RX_PKT_BUF_SZ,
79 DMA_FROM_DEVICE);
80 rh = (struct rx_header *) skb->data;
81 @@ -756,7 +756,7 @@
82 skb_size = b44_alloc_rx_skb(bp, cons, bp->rx_prod);
83 if (skb_size < 0)
84 goto drop_it;
85 - dma_unmap_single(&bp->sdev->dev, map,
86 + dma_unmap_single(bp->sdev->dev, map,
87 skb_size, DMA_FROM_DEVICE);
88 /* Leave out rx_header */
89 skb_put(skb, len + RX_PKT_OFFSET);
90 @@ -928,23 +928,23 @@
91 goto err_out;
92 }
93
94 - mapping = dma_map_single(&bp->sdev->dev, skb->data, len, DMA_TO_DEVICE);
95 + mapping = dma_map_single(bp->sdev->dev, skb->data, len, DMA_TO_DEVICE);
96 if (dma_mapping_error(mapping) || mapping + len > DMA_30BIT_MASK) {
97 struct sk_buff *bounce_skb;
98
99 /* Chip can't handle DMA to/from >1GB, use bounce buffer */
100 if (!dma_mapping_error(mapping))
101 - dma_unmap_single(&bp->sdev->dev, mapping, len, DMA_TO_DEVICE);
102 + dma_unmap_single(bp->sdev->dev, mapping, len, DMA_TO_DEVICE);
103
104 bounce_skb = __dev_alloc_skb(len, GFP_ATOMIC | GFP_DMA);
105 if (!bounce_skb)
106 goto err_out;
107
108 - mapping = dma_map_single(&bp->sdev->dev, bounce_skb->data,
109 + mapping = dma_map_single(bp->sdev->dev, bounce_skb->data,
110 len, DMA_TO_DEVICE);
111 if (dma_mapping_error(mapping) || mapping + len > DMA_30BIT_MASK) {
112 if (!dma_mapping_error(mapping))
113 - dma_unmap_single(&bp->sdev->dev, mapping,
114 + dma_unmap_single(bp->sdev->dev, mapping,
115 len, DMA_TO_DEVICE);
116 dev_kfree_skb_any(bounce_skb);
117 goto err_out;
118 @@ -1043,7 +1043,7 @@
119
120 if (rp->skb == NULL)
121 continue;
122 - dma_unmap_single(&bp->sdev->dev,
123 + dma_unmap_single(bp->sdev->dev,
124 pci_unmap_addr(rp, mapping),
125 RX_PKT_BUF_SZ,
126 DMA_FROM_DEVICE);
127 @@ -1057,7 +1057,7 @@
128
129 if (rp->skb == NULL)
130 continue;
131 - dma_unmap_single(&bp->sdev->dev,
132 + dma_unmap_single(bp->sdev->dev,
133 pci_unmap_addr(rp, mapping),
134 rp->skb->len,
135 DMA_TO_DEVICE);
136 @@ -1082,12 +1082,12 @@
137 memset(bp->tx_ring, 0, B44_TX_RING_BYTES);
138
139 if (bp->flags & B44_FLAG_RX_RING_HACK)
140 - dma_sync_single_for_device(&bp->sdev->dev, bp->rx_ring_dma,
141 + dma_sync_single_for_device(bp->sdev->dev, bp->rx_ring_dma,
142 DMA_TABLE_BYTES,
143 DMA_BIDIRECTIONAL);
144
145 if (bp->flags & B44_FLAG_TX_RING_HACK)
146 - dma_sync_single_for_device(&bp->sdev->dev, bp->tx_ring_dma,
147 + dma_sync_single_for_device(bp->sdev->dev, bp->tx_ring_dma,
148 DMA_TABLE_BYTES,
149 DMA_TO_DEVICE);
150
151 @@ -1109,24 +1109,24 @@
152 bp->tx_buffers = NULL;
153 if (bp->rx_ring) {
154 if (bp->flags & B44_FLAG_RX_RING_HACK) {
155 - dma_unmap_single(&bp->sdev->dev, bp->rx_ring_dma,
156 + dma_unmap_single(bp->sdev->dev, bp->rx_ring_dma,
157 DMA_TABLE_BYTES,
158 DMA_BIDIRECTIONAL);
159 kfree(bp->rx_ring);
160 } else
161 - dma_free_coherent(&bp->sdev->dev, DMA_TABLE_BYTES,
162 + dma_free_coherent(bp->sdev->dev, DMA_TABLE_BYTES,
163 bp->rx_ring, bp->rx_ring_dma);
164 bp->rx_ring = NULL;
165 bp->flags &= ~B44_FLAG_RX_RING_HACK;
166 }
167 if (bp->tx_ring) {
168 if (bp->flags & B44_FLAG_TX_RING_HACK) {
169 - dma_unmap_single(&bp->sdev->dev, bp->tx_ring_dma,
170 + dma_unmap_single(bp->sdev->dev, bp->tx_ring_dma,
171 DMA_TABLE_BYTES,
172 DMA_TO_DEVICE);
173 kfree(bp->tx_ring);
174 } else
175 - dma_free_coherent(&bp->sdev->dev, DMA_TABLE_BYTES,
176 + dma_free_coherent(bp->sdev->dev, DMA_TABLE_BYTES,
177 bp->tx_ring, bp->tx_ring_dma);
178 bp->tx_ring = NULL;
179 bp->flags &= ~B44_FLAG_TX_RING_HACK;
180 @@ -1152,7 +1152,7 @@
181 goto out_err;
182
183 size = DMA_TABLE_BYTES;
184 - bp->rx_ring = dma_alloc_coherent(&bp->sdev->dev, size, &bp->rx_ring_dma, GFP_ATOMIC);
185 + bp->rx_ring = dma_alloc_coherent(bp->sdev->dev, size, &bp->rx_ring_dma, GFP_ATOMIC);
186 if (!bp->rx_ring) {
187 /* Allocation may have failed due to pci_alloc_consistent
188 insisting on use of GFP_DMA, which is more restrictive
189 @@ -1164,7 +1164,7 @@
190 if (!rx_ring)
191 goto out_err;
192
193 - rx_ring_dma = dma_map_single(&bp->sdev->dev, rx_ring,
194 + rx_ring_dma = dma_map_single(bp->sdev->dev, rx_ring,
195 DMA_TABLE_BYTES,
196 DMA_BIDIRECTIONAL);
197
198 @@ -1179,7 +1179,7 @@
199 bp->flags |= B44_FLAG_RX_RING_HACK;
200 }
201
202 - bp->tx_ring = dma_alloc_coherent(&bp->sdev->dev, size, &bp->tx_ring_dma, GFP_ATOMIC);
203 + bp->tx_ring = dma_alloc_coherent(bp->sdev->dev, size, &bp->tx_ring_dma, GFP_ATOMIC);
204 if (!bp->tx_ring) {
205 /* Allocation may have failed due to dma_alloc_coherent
206 insisting on use of GFP_DMA, which is more restrictive
207 @@ -1191,7 +1191,7 @@
208 if (!tx_ring)
209 goto out_err;
210
211 - tx_ring_dma = dma_map_single(&bp->sdev->dev, tx_ring,
212 + tx_ring_dma = dma_map_single(bp->sdev->dev, tx_ring,
213 DMA_TABLE_BYTES,
214 DMA_TO_DEVICE);
215
216 @@ -2288,7 +2288,7 @@
217 bp->flags |= B44_FLAG_BUGGY_TXPTR;
218 */
219
220 - if (bp->sdev->dev->id.revision >= 7)
221 + if (bp->sdev->id.revision >= 7)
222 bp->flags |= B44_FLAG_B0_ANDLATER;
223
224 return err;
225 @@ -2298,7 +2298,6 @@
226 const struct ssb_device_id *ent)
227 {
228 static int b44_version_printed = 0;
229 - unsigned long b44reg_base, b44reg_len;
230 struct net_device *dev;
231 struct b44 *bp;
232 int err, i;
233 @@ -2310,13 +2309,13 @@
234
235 dev = alloc_etherdev(sizeof(*bp));
236 if (!dev) {
237 - dev_err(&pdev->dev, "Etherdev alloc failed, aborting.\n");
238 + dev_err(sdev->dev, "Etherdev alloc failed, aborting.\n");
239 err = -ENOMEM;
240 goto out;
241 }
242
243 SET_MODULE_OWNER(dev);
244 - SET_NETDEV_DEV(dev,&sdev->dev);
245 + SET_NETDEV_DEV(dev,sdev->dev);
246
247 /* No interesting netdevice features in this card... */
248 dev->features |= 0;
249 @@ -2354,7 +2353,7 @@
250
251 err = b44_get_invariants(bp);
252 if (err) {
253 - dev_err(&sdev->dev,
254 + dev_err(sdev->dev,
255 "Problem fetching invariants of chip, aborting.\n");
256 goto err_out_free_dev;
257 }
258 @@ -2375,7 +2374,7 @@
259
260 err = register_netdev(dev);
261 if (err) {
262 - dev_err(&sdev->dev, "Cannot register net device, aborting.\n");
263 + dev_err(sdev->dev, "Cannot register net device, aborting.\n");
264 goto out;
265 }
266
267 @@ -2454,7 +2453,6 @@
268 rc = request_irq(dev->irq, b44_interrupt, IRQF_SHARED, dev->name, dev);
269 if (rc) {
270 printk(KERN_ERR PFX "%s: request_irq failed\n", dev->name);
271 - pci_disable_device(pdev);
272 return rc;
273 }
274