|
@@ -287,6 +287,8 @@ int snd_emu10k1_memblk_map(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *b
|
|
return err;
|
|
return err;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+EXPORT_SYMBOL(snd_emu10k1_memblk_map);
|
|
|
|
+
|
|
/*
|
|
/*
|
|
* page allocation for DMA
|
|
* page allocation for DMA
|
|
*/
|
|
*/
|
|
@@ -387,6 +389,7 @@ snd_emu10k1_synth_alloc(struct snd_emu10k1 *hw, unsigned int size)
|
|
return (struct snd_util_memblk *)blk;
|
|
return (struct snd_util_memblk *)blk;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+EXPORT_SYMBOL(snd_emu10k1_synth_alloc);
|
|
|
|
|
|
/*
|
|
/*
|
|
* free a synth sample area
|
|
* free a synth sample area
|
|
@@ -409,6 +412,7 @@ snd_emu10k1_synth_free(struct snd_emu10k1 *emu, struct snd_util_memblk *memblk)
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+EXPORT_SYMBOL(snd_emu10k1_synth_free);
|
|
|
|
|
|
/* check new allocation range */
|
|
/* check new allocation range */
|
|
static void get_single_page_range(struct snd_util_memhdr *hdr,
|
|
static void get_single_page_range(struct snd_util_memhdr *hdr,
|
|
@@ -540,6 +544,8 @@ int snd_emu10k1_synth_bzero(struct snd_emu10k1 *emu, struct snd_util_memblk *blk
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+EXPORT_SYMBOL(snd_emu10k1_synth_bzero);
|
|
|
|
+
|
|
/*
|
|
/*
|
|
* copy_from_user(blk + offset, data, size)
|
|
* copy_from_user(blk + offset, data, size)
|
|
*/
|
|
*/
|
|
@@ -568,3 +574,5 @@ int snd_emu10k1_synth_copy_from_user(struct snd_emu10k1 *emu, struct snd_util_me
|
|
} while (offset < end_offset);
|
|
} while (offset < end_offset);
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
+
|
|
|
|
+EXPORT_SYMBOL(snd_emu10k1_synth_copy_from_user);
|