Revision c227f099 hw/etraxfs_dma.c
b/hw/etraxfs_dma.c | ||
---|---|---|
211 | 211 |
&& ctrl->channels[c].client; |
212 | 212 |
} |
213 | 213 |
|
214 |
static inline int fs_channel(a_target_phys_addr addr)
|
|
214 |
static inline int fs_channel(target_phys_addr_t addr)
|
|
215 | 215 |
{ |
216 | 216 |
/* Every channel has a 0x2000 ctrl register map. */ |
217 | 217 |
return addr >> 13; |
... | ... | |
220 | 220 |
#ifdef USE_THIS_DEAD_CODE |
221 | 221 |
static void channel_load_g(struct fs_dma_ctrl *ctrl, int c) |
222 | 222 |
{ |
223 |
a_target_phys_addr addr = channel_reg(ctrl, c, RW_GROUP);
|
|
223 |
target_phys_addr_t addr = channel_reg(ctrl, c, RW_GROUP);
|
|
224 | 224 |
|
225 | 225 |
/* Load and decode. FIXME: handle endianness. */ |
226 | 226 |
cpu_physical_memory_read (addr, |
... | ... | |
252 | 252 |
|
253 | 253 |
static void channel_load_c(struct fs_dma_ctrl *ctrl, int c) |
254 | 254 |
{ |
255 |
a_target_phys_addr addr = channel_reg(ctrl, c, RW_GROUP_DOWN);
|
|
255 |
target_phys_addr_t addr = channel_reg(ctrl, c, RW_GROUP_DOWN);
|
|
256 | 256 |
|
257 | 257 |
/* Load and decode. FIXME: handle endianness. */ |
258 | 258 |
cpu_physical_memory_read (addr, |
... | ... | |
269 | 269 |
|
270 | 270 |
static void channel_load_d(struct fs_dma_ctrl *ctrl, int c) |
271 | 271 |
{ |
272 |
a_target_phys_addr addr = channel_reg(ctrl, c, RW_SAVED_DATA);
|
|
272 |
target_phys_addr_t addr = channel_reg(ctrl, c, RW_SAVED_DATA);
|
|
273 | 273 |
|
274 | 274 |
/* Load and decode. FIXME: handle endianness. */ |
275 | 275 |
D(printf("%s ch=%d addr=" TARGET_FMT_plx "\n", __func__, c, addr)); |
... | ... | |
283 | 283 |
|
284 | 284 |
static void channel_store_c(struct fs_dma_ctrl *ctrl, int c) |
285 | 285 |
{ |
286 |
a_target_phys_addr addr = channel_reg(ctrl, c, RW_GROUP_DOWN);
|
|
286 |
target_phys_addr_t addr = channel_reg(ctrl, c, RW_GROUP_DOWN);
|
|
287 | 287 |
|
288 | 288 |
/* Encode and store. FIXME: handle endianness. */ |
289 | 289 |
D(printf("%s ch=%d addr=" TARGET_FMT_plx "\n", __func__, c, addr)); |
... | ... | |
295 | 295 |
|
296 | 296 |
static void channel_store_d(struct fs_dma_ctrl *ctrl, int c) |
297 | 297 |
{ |
298 |
a_target_phys_addr addr = channel_reg(ctrl, c, RW_SAVED_DATA);
|
|
298 |
target_phys_addr_t addr = channel_reg(ctrl, c, RW_SAVED_DATA);
|
|
299 | 299 |
|
300 | 300 |
/* Encode and store. FIXME: handle endianness. */ |
301 | 301 |
D(printf("%s ch=%d addr=" TARGET_FMT_plx "\n", __func__, c, addr)); |
... | ... | |
556 | 556 |
return 0; |
557 | 557 |
} |
558 | 558 |
|
559 |
static uint32_t dma_rinvalid (void *opaque, a_target_phys_addr addr)
|
|
559 |
static uint32_t dma_rinvalid (void *opaque, target_phys_addr_t addr)
|
|
560 | 560 |
{ |
561 | 561 |
hw_error("Unsupported short raccess. reg=" TARGET_FMT_plx "\n", addr); |
562 | 562 |
return 0; |
563 | 563 |
} |
564 | 564 |
|
565 | 565 |
static uint32_t |
566 |
dma_readl (void *opaque, a_target_phys_addr addr)
|
|
566 |
dma_readl (void *opaque, target_phys_addr_t addr)
|
|
567 | 567 |
{ |
568 | 568 |
struct fs_dma_ctrl *ctrl = opaque; |
569 | 569 |
int c; |
... | ... | |
591 | 591 |
} |
592 | 592 |
|
593 | 593 |
static void |
594 |
dma_winvalid (void *opaque, a_target_phys_addr addr, uint32_t value)
|
|
594 |
dma_winvalid (void *opaque, target_phys_addr_t addr, uint32_t value)
|
|
595 | 595 |
{ |
596 | 596 |
hw_error("Unsupported short waccess. reg=" TARGET_FMT_plx "\n", addr); |
597 | 597 |
} |
... | ... | |
608 | 608 |
} |
609 | 609 |
|
610 | 610 |
static void |
611 |
dma_writel (void *opaque, a_target_phys_addr addr, uint32_t value)
|
|
611 |
dma_writel (void *opaque, target_phys_addr_t addr, uint32_t value)
|
|
612 | 612 |
{ |
613 | 613 |
struct fs_dma_ctrl *ctrl = opaque; |
614 | 614 |
int c; |
... | ... | |
739 | 739 |
qemu_bh_schedule_idle(etraxfs_dmac->bh); |
740 | 740 |
} |
741 | 741 |
|
742 |
void *etraxfs_dmac_init(a_target_phys_addr base, int nr_channels)
|
|
742 |
void *etraxfs_dmac_init(target_phys_addr_t base, int nr_channels)
|
|
743 | 743 |
{ |
744 | 744 |
struct fs_dma_ctrl *ctrl = NULL; |
745 | 745 |
|
Also available in: Unified diff