`
弄月吟风
  • 浏览: 196888 次
  • 性别: Icon_minigender_1
  • 来自: 杭州
社区版块
存档分类
最新评论

Android语音采集

阅读更多

Android端的语音采集主要是调用AudioRecord,首先说几个参数

 

private static AudioRecord mRecord;
	// 音频获取源
	private int audioSource = MediaRecorder.AudioSource.MIC;
	// 设置音频采样率,44100是目前的标准,但是某些设备仍然支持22050,16000,11025
	private static int sampleRateInHz = 8000;// 44100;
	// 设置音频的录制的声道CHANNEL_IN_STEREO为双声道,CHANNEL_CONFIGURATION_MONO为单声道
	private static int channelConfig = AudioFormat.CHANNEL_CONFIGURATION_MONO;// AudioFormat.CHANNEL_IN_STEREO;
	// 音频数据格式:PCM 16位每个样本。保证设备支持。PCM 8位每个样本。不一定能得到设备支持。
	private static int audioFormat = AudioFormat.ENCODING_PCM_16BIT;
	// 音频大小
	private int bufSize;

 然后初始化一下AudioRecord,过程如下:

 

 

bufSize = AudioRecord.getMinBufferSize(sampleRateInHz, channelConfig,
				audioFormat);
		mRecord = new AudioRecord(audioSource, sampleRateInHz, channelConfig,
				audioFormat, bufSize);

 初始化完毕以后就需要采集音频数据了:

 

mRecord.startRecording();
		short audiodata[] = new short[bufSize];
		while (isRecord) {
			int readsize = 0;
			while (isRecord == true) {
				readsize = mRecord.read(audiodata, 0, bufSize);
				try {
					for (int i = 0; i < readsize; i++) {
						//dout.writeShort(audiodata[i]);
                                              //数据处理
					}
				} catch (IOException e) {
					// TODO Auto-generated catch block
					e.printStackTrace();
				}
			}
		}
		mRecord.stop();
		audiodata = null;

 接下来是一个语音的播放了,我们这边不放的是采集到的语音流,即PCM无损格式的语音数据,如下:

参数:

 

	private static AudioTrack mTrack;
	// 音频类型
	private int streamType = AudioManager.STREAM_MUSIC;
	// 设置音频采样率,44100是目前的标准,但是某些设备仍然支持22050,16000,11025
	private int sampleRateInHz = 8000;// 44100;
	// 设置音频的录制的声道CHANNEL_IN_STEREO为双声道,CHANNEL_CONFIGURATION_MONO为单声道
	private int channelConfig = AudioFormat.CHANNEL_CONFIGURATION_MONO;// AudioFormat.CHANNEL_IN_STEREO;
	// 音频数据格式:PCM 16位每个样本。保证设备支持。PCM 8位每个样本。不一定能得到设备支持。
	private int audioFormat = AudioFormat.ENCODING_PCM_16BIT;
	// 音频大小
	private int bufSize;
	// 音频模式
	private int mode = AudioTrack.MODE_STREAM;
	protected boolean keepRuning = true;

 然后初始化播放器:

 

bufSize = AudioTrack.getMinBufferSize(sampleRateInHz, channelConfig,
				audioFormat);
		mTrack = new AudioTrack(streamType, sampleRateInHz, channelConfig,
				audioFormat, bufSize, mode);

 然后是播放:

 

DataOutputStream dos = null;
		mTrack.play();
		try {
			revSocket = server.accept();
			dos = new DataOutputStream(new BufferedOutputStream(
					new FileOutputStream(audioFile)));
			din = new DataInputStream(revSocket.getInputStream());
		} catch (IOException e) {
			// TODO Auto-generated catch block
			e.printStackTrace();
		}

		while (keepRuning) {
			short[] buffer = new short[bufSize / 4];
			try {
				Log.i("状态", "接收数据");
				for (int i = 0; din.available() > 0 && i < buffer.length; i++) {
					buffer[i] = din.readShort();
					dos.writeShort(buffer[i]);
					Log.i("状态", "接收数据," + String.valueOf(i));
				}
				short[] bytes_pkg = buffer.clone();
				mTrack.write(bytes_pkg, 0, bytes_pkg.length);
			} catch (Exception e) {
				e.printStackTrace();
			}
		}
		mTrack.stop();
		try {
			dos.close();
			din.close();
		} catch (IOException e) {
			// TODO Auto-generated catch block
			e.printStackTrace();
		}

	}
将语音数据保存到文件,并且将裸数据文件保存成可播放的WAV文件
/**
	 * 这里将数据写入文件,但是并不能播放,因为AudioRecord获得的音频是原始的裸音频,
	 * 如果需要播放就必须加入一些格式或者编码的头信息。但是这样的好处就是你可以对音频的 裸数据进行处理,比如你要做一个爱说话的TOM
	 * 猫在这里就进行音频的处理,然后重新封装 所以说这样得到的音频比较容易做一些音频的处理。
	 */
	private void writeDateTOFile() {
		// new一个byte数组用来存一些字节数据,大小为缓冲区大小
		byte[] audiodata = new byte[minBufSize];
		FileOutputStream fos = null;
		int readsize = 0;
		try {
			File file = new File(AudioName);
			if (file.exists()) {
				file.delete();
			}
			fos = new FileOutputStream(file);// 建立一个可存取字节的文件
		} catch (Exception e) {
			e.printStackTrace();
		}
		while (isRecord == true) {
			readsize = mRecord.read(audiodata, 0, minBufSize);
			Log.i("采集大小", String.valueOf(readsize));
			if (AudioRecord.ERROR_INVALID_OPERATION != readsize) {
				try {
					fos.write(audiodata);
				} catch (IOException e) {
					e.printStackTrace();
				}
			}
		}
		try {
			fos.close();// 关闭写入流
		} catch (IOException e) {
			e.printStackTrace();
		}
	}

	// 这里得到可播放的音频文件
	private void copyWaveFile(String inFilename, String outFilename) {
		FileInputStream in = null;
		FileOutputStream out = null;
		long totalAudioLen = 0;
		long totalDataLen = totalAudioLen + 36;
		long longSampleRate = sampleRateInHz;
		int channels = 2;
		long byteRate = 16 * sampleRateInHz * channels / 8;
		byte[] data = new byte[minBufSize];
		try {
			in = new FileInputStream(inFilename);
			out = new FileOutputStream(outFilename);
			totalAudioLen = in.getChannel().size();
			totalDataLen = totalAudioLen + 36;
			WriteWaveFileHeader(out, totalAudioLen, totalDataLen,
					longSampleRate, channels, byteRate);
			while (in.read(data) != -1) {
				out.write(data);
			}
			in.close();
			out.close();
		} catch (FileNotFoundException e) {
			e.printStackTrace();
		} catch (IOException e) {
			e.printStackTrace();
		}
	}

	/**
	 * 这里提供一个头信息。插入这些信息就可以得到可以播放的文件。 为我为啥插入这44个字节,这个还真没深入研究,不过你随便打开一个wav
	 * 音频的文件,可以发现前面的头文件可以说基本一样哦。每种格式的文件都有 自己特有的头文件。
	 */
	private void WriteWaveFileHeader(FileOutputStream out, long totalAudioLen,
			long totalDataLen, long longSampleRate, int channels, long byteRate)
			throws IOException {
		byte[] header = new byte[44];
		header[0] = 'R'; // RIFF/WAVE header
		header[1] = 'I';
		header[2] = 'F';
		header[3] = 'F';
		header[4] = (byte) (totalDataLen & 0xff);
		header[5] = (byte) ((totalDataLen >> 8) & 0xff);
		header[6] = (byte) ((totalDataLen >> 16) & 0xff);
		header[7] = (byte) ((totalDataLen >> 24) & 0xff);
		header[8] = 'W';
		header[9] = 'A';
		header[10] = 'V';
		header[11] = 'E';
		header[12] = 'f'; // 'fmt ' chunk
		header[13] = 'm';
		header[14] = 't';
		header[15] = ' ';
		header[16] = 16; // 4 bytes: size of 'fmt ' chunk
		header[17] = 0;
		header[18] = 0;
		header[19] = 0;
		header[20] = 1; // format = 1
		header[21] = 0;
		header[22] = (byte) channels;
		header[23] = 0;
		header[24] = (byte) (longSampleRate & 0xff);
		header[25] = (byte) ((longSampleRate >> 8) & 0xff);
		header[26] = (byte) ((longSampleRate >> 16) & 0xff);
		header[27] = (byte) ((longSampleRate >> 24) & 0xff);
		header[28] = (byte) (byteRate & 0xff);
		header[29] = (byte) ((byteRate >> 8) & 0xff);
		header[30] = (byte) ((byteRate >> 16) & 0xff);
		header[31] = (byte) ((byteRate >> 24) & 0xff);
		header[32] = (byte) (2 * 16 / 8); // block align
		header[33] = 0;
		header[34] = 16; // bits per sample
		header[35] = 0;
		header[36] = 'd';
		header[37] = 'a';
		header[38] = 't';
		header[39] = 'a';
		header[40] = (byte) (totalAudioLen & 0xff);
		header[41] = (byte) ((totalAudioLen >> 8) & 0xff);
		header[42] = (byte) ((totalAudioLen >> 16) & 0xff);
		header[43] = (byte) ((totalAudioLen >> 24) & 0xff);
		out.write(header, 0, 44);
	}
 播放裸语音数据文件
short[] buffer = new short[bufferSize / 4];
			try {
				// 定义输入流,将音频写入到AudioTrack类中,实现播放
				DataInputStream dis = new DataInputStream(
						new BufferedInputStream(new FileInputStream(audioFile)));
				// 实例AudioTrack
				AudioTrack track = new AudioTrack(AudioManager.STREAM_MUSIC,
						frequence, channelConfig, audioEncoding, bufferSize,
						AudioTrack.MODE_STREAM);
				// 开始播放
				track.play();
				// 由于AudioTrack播放的是流,所以,我们需要一边播放一边读取
				while (isPlaying && dis.available() > 0) {
					int i = 0;
					while (dis.available() > 0 && i < buffer.length) {
						buffer[i] = dis.readShort();
						i++;
					}
					// 然后将数据写入到AudioTrack中
					track.write(buffer, 0, buffer.length);

				}

				// 播放结束
				track.stop();
				dis.close();
			} catch (Exception e) {
				// TODO: handle exception
			}
 
分享到:
评论

相关推荐

Global site tag (gtag.js) - Google Analytics