All Downloads are FREE. Search and download functionalities are using the official Maven repository.

jogamp.opengl.android.av.AndroidGLMediaPlayerAPI14 Maven / Gradle / Ivy

/**
 * Copyright 2012 JogAmp Community. All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without modification, are
 * permitted provided that the following conditions are met:
 * 
 *    1. Redistributions of source code must retain the above copyright notice, this list of
 *       conditions and the following disclaimer.
 * 
 *    2. Redistributions in binary form must reproduce the above copyright notice, this list
 *       of conditions and the following disclaimer in the documentation and/or other materials
 *       provided with the distribution.
 * 
 * THIS SOFTWARE IS PROVIDED BY JogAmp Community ``AS IS'' AND ANY EXPRESS OR IMPLIED
 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND
 * FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL JogAmp Community OR
 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
 * ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 * 
 * The views and conclusions contained in the software and documentation are those of the
 * authors and should not be interpreted as representing official policies, either expressed
 * or implied, of JogAmp Community.
 */
package jogamp.opengl.android.av;

import java.io.IOException;
import java.util.List;

import javax.media.opengl.GL;
import javax.media.opengl.GLES2;
import javax.media.opengl.GLException;

import com.jogamp.common.os.AndroidVersion;
import com.jogamp.common.os.Platform;
import com.jogamp.opengl.util.TimeFrameI;
import com.jogamp.opengl.util.av.GLMediaPlayer;
import com.jogamp.opengl.util.texture.Texture;
import com.jogamp.opengl.util.texture.TextureSequence;

import jogamp.common.os.android.StaticContext;
import jogamp.opengl.util.av.GLMediaPlayerImpl;

import android.graphics.SurfaceTexture;
import android.graphics.SurfaceTexture.OnFrameAvailableListener;
import android.hardware.Camera;
import android.media.MediaPlayer;
import android.media.MediaPlayer.OnCompletionListener;
import android.net.Uri;
import android.view.Surface;

/***
 * Android implementation utilizes API level 14 (4.0.? ICS) features
 * as listed below.
 * 

* We utilize the {@link MediaPlayer} with direct to texture streaming. * The MediaPlayer uses libstagefright to access the OpenMAX AL implementation * for hardware decoding. *

*
    *
  • Android API Level 14: {@link MediaPlayer#setSurface(Surface)}
  • *
  • Android API Level 14: {@link Surface#Surface(android.graphics.SurfaceTexture)}
  • *
*

* Since the MediaPlayer API can only deal w/ one SurfaceTexture, * we enforce textureCount = 2 via {@link #validateTextureCount(int)} * and duplicate the single texture via {@link #createTexFrames(GL, int)} .. etc. * Two instanceds of TextureFrame are required due our framework implementation w/ Ringbuffer and 'lastFrame' access. *

*/ public class AndroidGLMediaPlayerAPI14 extends GLMediaPlayerImpl { static final boolean available; static { boolean _avail = false; if(Platform.OS_TYPE.equals(Platform.OSType.ANDROID)) { if(AndroidVersion.SDK_INT >= 14) { _avail = true; } } available = _avail; } public static final boolean isAvailable() { return available; } private MediaPlayer mp; private Camera cam; private long playStart = 0; private volatile boolean updateSurface = false; private Object updateSurfaceLock = new Object(); private SurfaceTextureFrame singleSTexFrame = null; private int sTexFrameCount = 0; private boolean sTexFrameAttached = false; private volatile boolean eos = false; /** private static String toString(MediaPlayer m) { if(null == m) return ""; return "MediaPlayer[playing "+m.isPlaying()+", pos "+m.getCurrentPosition()/1000.0f+"s, "+m.getVideoWidth()+"x"+m.getVideoHeight()+"]"; } */ public AndroidGLMediaPlayerAPI14() { super(); if(!available) { throw new RuntimeException("AndroidGLMediaPlayerAPI14 not available"); } this.setTextureTarget(GLES2.GL_TEXTURE_EXTERNAL_OES); } @Override protected final boolean setPlaySpeedImpl(float rate) { // FIXME return false; } @Override protected final boolean setAudioVolumeImpl(float v) { if(null != mp) { try { mp.setVolume(v, v); return true; } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } } return false; } @Override protected final boolean playImpl() { playStart = Platform.currentTimeMillis(); if(null != mp) { try { mp.start(); eos = false; mp.setOnCompletionListener(onCompletionListener); return true; } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } } else if( null != cam ) { try { if( sTexFrameAttached ) { cam.startPreview(); } return true; } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } } return false; } @Override protected final boolean pauseImpl() { if(null != mp) { wakeUp(false); try { mp.pause(); return true; } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } } else if( null != cam ) { wakeUp(false); try { cam.stopPreview(); return true; } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } } return false; } @Override protected final int seekImpl(int msec) { if(null != mp) { mp.seekTo(msec); return mp.getCurrentPosition(); } return 0; } private void wakeUp(boolean newFrame) { synchronized(updateSurfaceLock) { if(newFrame) { updateSurface = true; } updateSurfaceLock.notifyAll(); } } @Override protected final int getAudioPTSImpl() { return null != mp ? mp.getCurrentPosition() : 0; } @Override protected final void destroyImpl(GL gl) { if(null != mp) { wakeUp(false); try { mp.stop(); } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } mp.release(); mp = null; } if( null != cam ) { wakeUp(false); try { cam.stopPreview(); } catch (IllegalStateException ise) { if(DEBUG) { ise.printStackTrace(); } } cam.release(); cam = null; } } public static class SurfaceTextureFrame extends TextureSequence.TextureFrame { public SurfaceTextureFrame(Texture t, SurfaceTexture stex) { super(t); this.surfaceTex = stex; } public String toString() { return "SurfaceTextureFrame[pts " + pts + " ms, l " + duration + " ms, texID "+ texture.getTextureObject() + ", " + surfaceTex + "]"; } public final SurfaceTexture surfaceTex; } @Override protected final void initStreamImpl(int vid, int aid) throws IOException { if( null == streamLoc ) { return; } if( null == mp && null == cam ) { if( null == cameraPath ) { mp = new MediaPlayer(); } else { int cameraId = 0; try { cameraId = Integer.valueOf(cameraPath); } catch (NumberFormatException nfe) {} if( 0 <= cameraId && cameraId < Camera.getNumberOfCameras() ) { cam = Camera.open(cameraId); } else { cam = Camera.open(); } } } if(null!=mp) { if( GLMediaPlayer.STREAM_ID_NONE == aid ) { mp.setVolume(0f, 0f); // FIXME: Disable audio handling } // else FIXME: Select aid ! // Note: Both FIXMEs seem to be n/a via Android's MediaPlayer -> Switch to API level 16 MediaCodec/MediaExtractor .. try { final Uri _uri = Uri.parse(streamLoc.toString()); mp.setDataSource(StaticContext.getContext(), _uri); } catch (IllegalArgumentException e) { throw new RuntimeException(e); } catch (SecurityException e) { throw new RuntimeException(e); } catch (IllegalStateException e) { throw new RuntimeException(e); } mp.setSurface(null); try { mp.prepare(); } catch (IOException ioe) { throw new IOException("MediaPlayer failed to process stream <"+streamLoc.toString()+">: "+ioe.getMessage(), ioe); } final int r_aid = GLMediaPlayer.STREAM_ID_NONE == aid ? GLMediaPlayer.STREAM_ID_NONE : 1 /* fake */; final String icodec = "android"; updateAttributes(0 /* fake */, r_aid, mp.getVideoWidth(), mp.getVideoHeight(), 0, 0, 0, 0f, 0, 0, mp.getDuration(), icodec, icodec); } else if( null != cam ) { final String icodec = "android"; final int[] fpsRange = { 0, 0 }; final Camera.Parameters p = cam.getParameters(); p.getPreviewFpsRange(fpsRange); final Camera.Size size = p.getPreviewSize(); if( DEBUG ) { final int picFmt = p.getPictureFormat(); final Camera.Size prefSize = p.getPreferredPreviewSizeForVideo(); System.err.println("MediaPlayer.Camera: fps "+fpsRange[0]+".."+fpsRange[1]+", size[pref "+camSz2Str(prefSize)+", cur "+camSz2Str(size)+"], fmt "+picFmt); List supSizes = p.getSupportedVideoSizes(); for(int i=0; i * Returns 2 - implementation duplicates single texture *

*/ @Override protected int validateTextureCount(int desiredTextureCount) { return 2; } @Override protected final int getNextTextureImpl(GL gl, TextureFrame nextFrame) { int pts = TimeFrameI.INVALID_PTS; if(null != mp || null != cam) { final SurfaceTextureFrame sTexFrame = (SurfaceTextureFrame) nextFrame; final SurfaceTexture surfTex = sTexFrame.surfaceTex; if( sTexFrame != singleSTexFrame ) { throw new InternalError("XXX: sTexFrame: "+sTexFrame+", singleSTexFrame "+singleSTexFrame); } if( !sTexFrameAttached ) { sTexFrameAttached = true; final Surface surface; if( null != mp ) { surface = new Surface(sTexFrame.surfaceTex); mp.setSurface(surface); } else { surface = null; try { cam.setPreviewTexture(sTexFrame.surfaceTex); cam.startPreview(); } catch (IOException ioe) { throw new RuntimeException("MediaPlayer failed to process stream <"+streamLoc.toString()+">: "+ioe.getMessage(), ioe); } } if( null != surface ) { surface.release(); } surfTex.setOnFrameAvailableListener(onFrameAvailableListener); } if( eos || (null != mp && !mp.isPlaying() ) ) { eos = true; pts = TimeFrameI.END_OF_STREAM_PTS; } else { // Only block once, no while-loop. // This relaxes locking code of non crucial resources/events. boolean update = updateSurface; if( !update ) { synchronized(updateSurfaceLock) { if(!updateSurface) { // volatile OK. try { updateSurfaceLock.wait(); } catch (InterruptedException e) { e.printStackTrace(); } } update = updateSurface; updateSurface = false; } } if(update) { surfTex.updateTexImage(); // nextFrame.setPTS( (int) ( nextSTex.getTimestamp() / 1000000L ) ); // nano -9 -> milli -3 if( null != mp ) { pts = mp.getCurrentPosition(); } else { pts = (int) ( Platform.currentTimeMillis() - playStart ); } // stex.getTransformMatrix(atex.getSTMatrix()); } } nextFrame.setPTS( pts ); } return pts; } /** * {@inheritDoc} *

* Creates only one single texture and duplicated content to 2 TextureFrames *

*/ @Override protected TextureFrame[] createTexFrames(GL gl, final int count) { final int[] texNames = new int[1]; gl.glGenTextures(1, texNames, 0); final int err = gl.glGetError(); if( GL.GL_NO_ERROR != err ) { throw new RuntimeException("TextureNames creation failed (num: 1/"+count+"): err "+toHexString(err)); } final TextureFrame[] texFrames = new TextureFrame[count]; for(int i=0; i * Returns the single texture, which is created at 1st call. *

*/ @Override protected final TextureSequence.TextureFrame createTexImage(GL gl, int texName) { sTexFrameCount++; if( 1 == sTexFrameCount ) { singleSTexFrame = new SurfaceTextureFrame( createTexImageImpl(gl, texName, width, height), new SurfaceTexture(texName) ); } return singleSTexFrame; } /** * {@inheritDoc} *

* Destroys the single texture at last call. *

*/ @Override protected final void destroyTexFrame(GL gl, TextureSequence.TextureFrame frame) { sTexFrameCount--; if( 0 == sTexFrameCount ) { singleSTexFrame = null; sTexFrameAttached = false; final SurfaceTextureFrame sFrame = (SurfaceTextureFrame) frame; sFrame.surfaceTex.release(); super.destroyTexFrame(gl, frame); } } private OnFrameAvailableListener onFrameAvailableListener = new OnFrameAvailableListener() { @Override public void onFrameAvailable(SurfaceTexture surfaceTexture) { wakeUp(true); } }; private OnCompletionListener onCompletionListener = new OnCompletionListener() { @Override public void onCompletion(MediaPlayer mp) { eos = true; } }; }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy