你怎么玩一个长的AudioClip?

我写了一个简单的类来在一个简单的游戏中播放音频文件。 它适用于枪声或爆炸等小声音,但当我尝试将它用于背景音乐时,我收到此错误:“无法分配剪辑数据:请求的缓冲区太大。” 我假设这意味着文件太大,但我怎么能解决这个问题呢? 资源:

import java.io.File; import javax.sound.sampled.AudioInputStream; import javax.sound.sampled.AudioSystem; import javax.sound.sampled.Clip; public class Sound{ private Clip clip; public Sound(String filepath){ System.out.println(filepath); File file = new File(filepath); try { clip = AudioSystem.getClip(); AudioInputStream inputStream = AudioSystem.getAudioInputStream(file); clip.open(inputStream); } catch (Exception e) { System.err.println(e.getMessage()); } } public void play(){ System.out.println("play"); if(clip.isActive()){ clip.stop(); } clip.setFramePosition(0); clip.start(); } public void stop(){ clip.stop(); } public void loop(){ if(!clip.isActive()){ clip.setFramePosition(0); clip.loop(Clip.LOOP_CONTINUOUSLY); }else{ System.out.println("ALREADY PLAYING"); } } public boolean getActive(){return clip.isActive();} } 

使用BigClip 。 这是我用来播放12-18分钟(或更多)的MP3的课程。

它需要运行时类路径上的mp3plugin.jar来实际加载MP3格式的声音,但这不是重点。 重点是:

  1. BigClip会将声音文件加载到JVM在OutOfMemoryError之前允许的最大内存。

 import java.awt.Component; import javax.swing.*; import javax.sound.sampled.*; import java.io.*; import java.util.logging.*; import java.util.Arrays; import java.net.URL; import javax.swing.JOptionPane; class BigClipExample { public static void main(String[] args) throws Exception { URL url = new URL("http://pscode.org/media/leftright.wav"); BigClip clip = new BigClip(); AudioInputStream ais = AudioSystem.getAudioInputStream(url); clip.open(ais); clip.start(); JOptionPane.showMessageDialog(null, "BigClip.start()"); clip.loop(4); JOptionPane.showMessageDialog(null, "BigClip.loop(4)"); clip.setFastForward(true); clip.loop(8); // the looping/FF combo. reveals a bug.. // there is a slight 'click' in the sound that should not be audible JOptionPane.showMessageDialog(null, "Are you on speed?"); } } /** An implementation of the javax.sound.sampled.Clip that is designed to handle Clips of arbitrary size, limited only by the amount of memory available to the app. It uses the post 1.4 thread behaviour (daemon thread) that will stop the sound running after the main has exited. 
  • 2012-02-29 - Reworked play/loop to fix several bugs.
  • 2009-09-01 - Fixed bug that had clip ..clipped at the end, by calling drain() (before calling stop()) on the dataline after the play loop was complete. Improvement to frame and microsecond position determination.
  • 2009-08-17 - added convenience constructor that accepts a Clip. Changed the private convertFrameToM..seconds methods from 'micro' to 'milli' to reflect that they were dealing with units of 1000/th of a second.
  • 2009-08-14 - got rid of flush() after the sound loop, as it was cutting off tracks just before the end, and was found to be not needed for the fast-forward/rewind functionality it was introduced to support.
  • 2009-08-11 - First binary release.
NB Remove @Override notation and logging to use in 1.3+ @since 1.5 @version 2012-02-29 @author Andrew Thompson @author Alejandro Garcia */ class BigClip implements Clip, LineListener { /** The DataLine used by this Clip. */ private SourceDataLine dataLine; /** The raw bytes of the audio data. */ private byte[] audioData; /** The stream wrapper for the audioData. */ private ByteArrayInputStream inputStream; /** Loop count set by the calling code. */ private int loopCount = 1; /** Internal count of how many loops to go. */ private int countDown = 1; /** The start of a loop point. Defaults to 0. */ private int loopPointStart; /** The end of a loop point. Defaults to the end of the Clip. */ private int loopPointEnd; /** Stores the current frame position of the clip. */ private int framePosition; /** Thread used to run() sound. */ private Thread thread; /** Whether the sound is currently playing or active. */ private boolean active; /** Stores the last time bytes were dumped to the audio stream. */ private long timelastPositionSet; private int bufferUpdateFactor = 2; /** The parent Component for the loading progress dialog. */ Component parent = null; /** Used for reporting messages. */ private Logger logger = Logger.getAnonymousLogger(); /** Default constructor for a BigClip. Does nothing. Information from the AudioInputStream passed in open() will be used to get an appropriate SourceDataLine. */ public BigClip() {} /** There are a number of AudioSystem methods that will return a configured Clip. This convenience constructor allows us to obtain a SourceDataLine for the BigClip that uses the same AudioFormat as the original Clip. @param clip Clip The Clip used to configure the BigClip. */ public BigClip(Clip clip) throws LineUnavailableException { dataLine = AudioSystem.getSourceDataLine( clip.getFormat() ); } /** Provides the entire audio buffer of this clip. @return audioData byte[] The bytes of the audio data that is loaded in this Clip. */ public byte[] getAudioData() { return audioData; } /** Sets a parent component to act as owner of a "Loading track.." progress dialog. If null, there will be no progress shown. */ public void setParentComponent(Component parent) { this.parent = parent; } /** Converts a frame count to a duration in milliseconds. */ private long convertFramesToMilliseconds(int frames) { return (frames/(long)dataLine.getFormat().getSampleRate())*1000; } /** Converts a duration in milliseconds to a frame count. */ private int convertMillisecondsToFrames(long milliseconds) { return (int)(milliseconds/dataLine.getFormat().getSampleRate()); } @Override public void update(LineEvent le) { logger.log(Level.FINEST, "update: " + le ); } @Override public void loop(int count) { logger.log(Level.FINEST, "loop(" + count + ") - framePosition: " + framePosition); loopCount = count; countDown = count; active = true; inputStream.reset(); start(); } @Override public void setLoopPoints(int start, int end) { if ( start<0 || start>audioData.length-1 || end<0 || end>audioData.length ) { throw new IllegalArgumentException( "Loop points '" + start + "' and '" + end + "' cannot be set for buffer of size " + audioData.length); } if (start>end) { throw new IllegalArgumentException( "End position " + end + " preceeds start position " + start); } loopPointStart = start; framePosition = loopPointStart; loopPointEnd = end; } @Override public void setMicrosecondPosition(long milliseconds) { framePosition = convertMillisecondsToFrames(milliseconds); } @Override public long getMicrosecondPosition() { return convertFramesToMilliseconds(getFramePosition()); } @Override public long getMicrosecondLength() { return convertFramesToMilliseconds(getFrameLength()); } @Override public void setFramePosition(int frames) { framePosition = frames; int offset = framePosition*format.getFrameSize(); try { inputStream.reset(); inputStream.read(new byte[offset]); } catch(Exception e) { e.printStackTrace(); } } @Override public int getFramePosition() { long timeSinceLastPositionSet = System.currentTimeMillis() - timelastPositionSet; int size = dataLine.getBufferSize()*(format.getChannels()/2)/bufferUpdateFactor; int framesSinceLast = (int)((timeSinceLastPositionSet/1000f)* dataLine.getFormat().getFrameRate()); int framesRemainingTillTime = size - framesSinceLast; return framePosition - framesRemainingTillTime; } @Override public int getFrameLength() { return audioData.length/format.getFrameSize(); } AudioFormat format; @Override public void open(AudioInputStream stream) throws IOException, LineUnavailableException { AudioInputStream is1; format = stream.getFormat(); if (format.getEncoding()!=AudioFormat.Encoding.PCM_SIGNED) { is1 = AudioSystem.getAudioInputStream( AudioFormat.Encoding.PCM_SIGNED, stream ); } else { is1 = stream; } format = is1.getFormat(); InputStream is2; if (parent!=null) { ProgressMonitorInputStream pmis = new ProgressMonitorInputStream( parent, "Loading track..", is1); pmis.getProgressMonitor().setMillisToPopup(0); is2 = pmis; } else { is2 = is1; } byte[] buf = new byte[ 2^16 ]; int totalRead = 0; int numRead = 0; ByteArrayOutputStream baos = new ByteArrayOutputStream(); numRead = is2.read( buf ); while (numRead>-1) { baos.write( buf, 0, numRead ); numRead = is2.read( buf, 0, buf.length ); totalRead += numRead; } is2.close(); audioData = baos.toByteArray(); AudioFormat afTemp; if (format.getChannels()<2) { afTemp = new AudioFormat( format.getEncoding(), format.getSampleRate(), format.getSampleSizeInBits(), 2, format.getSampleSizeInBits()*2/8, // calculate frame size format.getFrameRate(), format.isBigEndian() ); } else { afTemp = format; } setLoopPoints(0,audioData.length); dataLine = AudioSystem.getSourceDataLine(afTemp); dataLine.open(); inputStream = new ByteArrayInputStream( audioData ); } @Override public void open(AudioFormat format, byte[] data, int offset, int bufferSize) throws LineUnavailableException { byte[] input = new byte[bufferSize]; for (int ii=0; ii0) && active ) { logger.log(Level.FINEST, "BigClip.start() loop " + framePosition ); totalBytes += bytesRead; int framesRead; byte[] tempData; if (format.getChannels()<2) { tempData = convertMonoToStereo(data, bytesRead); framesRead = bytesRead/ format.getFrameSize(); bytesRead*=2; } else { framesRead = bytesRead/ dataLine.getFormat().getFrameSize(); tempData = Arrays.copyOfRange(data, 0, bytesRead); } framePosition += framesRead; if (framePosition>=loopPointEnd) { framePosition = loopPointStart; inputStream.reset(); countDown--; logger.log(Level.FINEST, "Loop Count: " + countDown ); } timelastPositionSet = System.currentTimeMillis(); byte[] newData; if (fastForward) { newData = getEveryNthFrame(tempData, 2); } else if (fastRewind) { byte[] temp = getEveryNthFrame(tempData, 2); newData = reverseFrames(temp); inputStream.reset(); totalBytes -= 2*bytesRead; framePosition -= 2*framesRead; if (totalBytes<0) { setFastRewind(false); totalBytes = 0; } inputStream.skip(totalBytes); logger.log(Level.FINE, "totalBytes " + totalBytes); } else { newData = tempData; } dataLine.write(newData, 0, newData.length); if (startOrMove) { data = new byte[bufSize/ bufferUpdateFactor]; startOrMove = false; } bytesRead = inputStream.read(data,0,data.length); if (bytesRead<0 && countDown-->1) { inputStream.read(new byte[offset], 0, offset); logger.log(Level.FINE, "loopCount " + loopCount ); logger.log(Level.FINE, "countDown " + countDown ); inputStream.reset(); bytesRead = inputStream.read(data,0,data.length); } } logger.log(Level.FINEST, "BigClip.start() loop ENDED" + framePosition ); active = false; countDown = 1; framePosition = 0; inputStream.reset(); dataLine.drain(); dataLine.stop(); /* should these open()/close() be here, or explicitly called by user program? */ dataLine.close(); } catch (LineUnavailableException lue) { logger.log( Level.SEVERE, "No sound line available!", lue ); if (parent!=null) { JOptionPane.showMessageDialog( parent, "Clear the sound lines to proceed", "No audio lines available!", JOptionPane.ERROR_MESSAGE); } } } }; thread= new Thread(r); // makes thread behaviour compatible with JavaSound post 1.4 thread.setDaemon(true); thread.start(); } /** Assume the frame size is 4. */ public byte[] reverseFrames(byte[] data) { byte[] reversed = new byte[data.length]; byte[] frame = new byte[4]; for (int ii=0; ii(data.length/4)-5) { logger.log(Level.FINER, "From \t" + first + " \tlast " + last ); logger.log(Level.FINER, "To \t" + ((ii*4)+0) + " \tlast " + ((ii*4)+3) ); } } /* for (int ii=0; iilargest) { largest = Math.abs(current); } } } else { for (int cc = 0; cc < samples; cc++) { current = (audioData[cc*2+1]*256 + (audioData[cc*2] & 0xFF)); if (Math.abs(current)>largest) { largest = Math.abs(current); } } } } else { for (int cc = 0; cc < samples; cc++) { current = (audioData[cc] & 0xFF); if (Math.abs(current)>largest) { largest = Math.abs(current); } } } } else { if (bitDepth/8==2) { if (bigEndian) { for (int cc = 0; cc < samples; cc++) { current = (audioData[cc*2]*256 + (audioData[cc*2+1] - 0x80)); if (Math.abs(current)>largest) { largest = Math.abs(current); } } } else { for (int cc = 0; cc < samples; cc++) { current = (audioData[cc*2+1]*256 + (audioData[cc*2] - 0x80)); if (Math.abs(current)>largest) { largest = Math.abs(current); } } } } else { for (int cc = 0; cc < samples; cc++) { if ( audioData[cc]>0 ) { current = (audioData[cc] - 0x80); if (Math.abs(current)>largest) { largest = Math.abs(current); } } else { current = (audioData[cc] + 0x80); if (Math.abs(current)>largest) { largest = Math.abs(current); } } } } } // audioData logger.log(Level.FINEST, "Max signal level: " + (double)largest/(Math.pow(2, bitDepth-1))); return (double)largest/(Math.pow(2, bitDepth-1)); } }

google让我来到这里: http : //docs.oracle.com/javase/tutorial/sound/sampled-overview.html
浏览完三个第一部分之后我就可以把它放在一起了:

  import javax.sound.sampled.*; import javax.sound.*; import java.io.*; public class Playme { Playme(String filename){ int total, totalToRead, numBytesRead, numBytesToRead; byte[] buffer; boolean stopped; AudioFormat wav; TargetDataLine line; SourceDataLine lineIn; DataLine.Info info; File file; FileInputStream fis; //AudioFormat(float sampleRate, int sampleSizeInBits, //int channels, boolean signed, boolean bigEndian) wav = new AudioFormat(44100, 16, 2, true, false); info = new DataLine.Info(SourceDataLine.class, wav); buffer = new byte[1024*333]; numBytesToRead = 1024*333; total=0; stopped = false; if (!AudioSystem.isLineSupported(info)) { System.out.print("no support for " + wav.toString() ); } try { // Obtain and open the line. lineIn = (SourceDataLine) AudioSystem.getLine(info); lineIn.open(wav); lineIn.start(); fis = new FileInputStream(file = new File(filename)); totalToRead = fis.available(); while (total < totalToRead && !stopped){ numBytesRead = fis.read(buffer, 0, numBytesToRead); if (numBytesRead == -1) break; total += numBytesRead; lineIn.write(buffer, 0, numBytesRead); } } catch (LineUnavailableException ex) { ex.printStackTrace(); } catch (FileNotFoundException nofile) { nofile.printStackTrace(); } catch (IOException io) { io.printStackTrace(); } } public static void main(String[] argv) { Playme mb_745 = new Playme(argv[0]); //Playme mb_745 = new Playme("/R/tmp/tmp/audiodump.wav"); } } 

请注意,可能存在以下错误:

 numBytesToRead = 1024*333; 

因为SourceDataLine.write的javadoc说:

 The number of bytes to write must represent an integral number of sample frames, such that: [ bytes written ] % [frame size in bytes ] == 0 

new AudioFormat(44100, 16, 2, true, false); 来自:

 $ file /R/tmp/tmp/audiodump.wav /R/tmp/tmp/audiodump.wav: RIFF (little-endian) data, WAVE audio, Microsoft PCM, 16 bit, stereo 44100 Hz 

所以现在我通过运行来听到这个巨大的745MB wav:

 javac Playme.java && java Playme /R/tmp/tmp/audiodump.wav 

希望你觉得这很有用,祝你好运!

我在建议的BigClip代码中修复了一些错误:修复了帧微秒的错误,反之亦然,并将私有convertFrameToM..seconds方法从’milli’改回’micro’。 现在getMicrosecondPosition()和setMicrosecondPosition()工作正确。现在getMicrosecondPosition(),setMicrosecondPosition(),getMicrosecondLength()工作正常。

 ....................... /** Converts a frame count to a duration in milliseconds. */ private long convertFramesToMicrosecond(int frames) { return (long)(frames / dataLine.getFormat().getSampleRate() * 1000000); } /** Converts a duration in milliseconds to a frame count. */ private int convertMicrosecondToFrames(long microsecond) { return (int) (microsecond / 1000000.0 * dataLine.getFormat().getSampleRate()); } ....................... @Override public void setMicrosecondPosition(long microsecond) { framePosition = convertMicrosecondToFrames(microsecond); } @Override public long getMicrosecondPosition() { return convertFramesToMicrosecond(getFramePosition()); } @Override public long getMicrosecondLength() { return convertFramesToMicrosecond(getFrameLength()); } .......................