I need to make an app that sends recorded audio in .wav format to a REST endpoint. I will be using AFNetworking for the POST, but I am not sure where to go with recording the audio and then converting it. I know I should use AVFoundation to record, but I am not sure how to convert. Would CoreAudio help?
Related
How can I save what AVPlayer is currently playing (both video and audio) from Live HLS stream?
I know how to load and play m3u8 video file using AVPlayer.
Please note that the HLS stream is live and not Video on demand so cannot use AVAggregateAssetDownloadTask In the perfect scenario will get CMSampleBuffer objects which can save to file easily. Also AVPlayerItemOutput is not entirely an option because I am unable to see how will get the audio channel.
Seems not possible with the current SDK. I've implemented it using ffmpeg
I have a remote MP4 file that has video and sound. I want to only stream the audio track of the mp4 file. I don't need the video and want to decrease the internet usage in my app. I have no idea where to start with this, Google doesn't seem to help. Is this impossible? Any ideas?
You can't... If you want audio only, you need to split video / audio on the server or put a stream software (like vlc, ffmpeg or mplayer) who can re-encode file in realtime (so you can drop video for new stream)
It's better, for me, to process all files on server and extract audio track...
I have an AVSpeechSynthesizer which converts text to speech, but i've encountered a problem.
I don't know how to save the audio file that it generates to a music file, which I would quite like to be able to do!
So here's my question, how do you save the AVSpeechSynthesizer output and if this isn't possible, can I us AVFoundation, CoreMedia or other public API to capture the output of the speakers, but before it has come out?
Thanks!
Unfortunately no, there is no public API available to capture the speaker output and looking over the docs for AVSpeechSynthesizer and related classes I don't see a way to capture any audio from it. You may want to look at 3rd party libraries to help with this.
Related questions:
Recording audio output only from speaker of iphone excluding microphone
Text-to-speech libraries for iPhone
I have NSDATA which is recorded through AudioQueue , I want to play this audio through any AUDIO player on IOS, Please help on this.
Thanks
Djrecker
Save it as an audio file (e.g. WAV). AudioFile and ExtAudioFile APIs can help you write this.
You can't expect any audio player on iOS to interpret/guess a blob of sample data correctly. This is why file formats exist -- audio file formats can specify sample count, channel count, sample rate, sample format, and a ton of other information about the audio in the file. Using an audio file, you can exchange this recording in a format other programs are able to read (correctly).
I want to play sounds from an internet server in my own program. But the sample codes Apple supplied concerned about sound play are all open an audio file, and then play it.
I want to know how can I play PCM data from memory, which received from internet continuously. Either OpenAL or AudioQueue is OK.
Give this a look:
http://cocoawithlove.com/2008/09/streaming-and-playing-live-mp3-stream.html
http://developer.apple.com/iphone/library/documentation/iPhone/Conceptual/iPhoneOSProgrammingGuide/AudioandVideoTechnologies/AudioandVideoTechnologies.html#//apple_ref/doc/uid/TP40007072-CH19-SW8