br.com.skylane.voicer.rtp.RtpMediaDecoder.java Source code

Java tutorial

Introduction

Here is the source code for br.com.skylane.voicer.rtp.RtpMediaDecoder.java

Source

/*
* Copyright (C) 2015 Creativa77 SRL and others
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* Contributors:
*
* Ayelen Chavez ashi@creativa77.com.ar
* Julian Cerruti jcerruti@creativa77.com.ar
*
*/

package br.com.skylane.voicer.rtp;

import java.net.DatagramPacket;
import java.nio.ByteBuffer;

import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.jboss.netty.buffer.ChannelBuffer;
import org.jboss.netty.buffer.ChannelBuffers;

import com.biasedbit.efflux.packet.DataPacket;

import android.media.MediaCodec;
import android.media.MediaCodecInfo;
import android.media.MediaFormat;
import android.view.Surface;
import android.view.SurfaceHolder;
import android.view.SurfaceView;
import br.com.skylane.voicer.VoicerHelper;
import br.com.skylane.voicer.udp.PacketReceivedListener;

/**
 * Implementation of the decoder that uses RTP as transport protocol to decode H264 encoded frames.
 * This object wraps up an Android API decoder and uses it to decode video frames.
 *
 * @author Ayelen Chavez
 */
public class RtpMediaDecoder implements SurfaceHolder.Callback, PacketReceivedListener {

    // configuration constants
    public static final String DEBUGGING_PROPERTY = "DEBUGGING";
    public static final String CONFIG_USE_NIO = "USE_NIO";
    public static final String CONFIG_BUFFER_TYPE = "BUFFER_TYPE";
    public static final String CONFIG_RECEIVE_BUFFER_SIZE = "RECEIVE_BUFFER_SIZE_BYTES";
    public static final int DATA_STREAMING_PORT = 5006;
    public static final int SURFACE_WIDTH = 640;
    public static final int SURFACE_HEIGHT = 480;
    public static final String TRANSPORT_PROTOCOL = "RTP";
    public static final String VIDEO_CODEC = "H.264";
    private static final int FRAME_RATE = 30; // 30fps
    private static final int IFRAME_INTERVAL = 5; // 5 seconds between I-frames

    // constant used to activate and deactivate logs
    public static boolean DEBUGGING = false;
    // surface view where to play video
    private final SurfaceView surfaceView;
    //private final Properties configuration;
    public String bufferType = "time-window";
    public boolean useNio = true;
    public int receiveBufferSize = 50000;
    private PlayerThread playerThread;
    private ByteBuffer[] inputBuffers;
    private ByteBuffer[] outputBuffers;
    private MediaCodec.BufferInfo info = new MediaCodec.BufferInfo();
    private MediaCodec decoder;
    private Log log = LogFactory.getLog(RtpMediaDecoder.class);
    private boolean currentFrameHasError = false;
    //private ByteArrayOutputStream baos;
    private ByteBuffer inputBuf;
    int inputBufIndex;
    private long currentTimestamp = 0;
    private int lastSequenceNumber = 0;
    private boolean lastSequenceNumberIsValid = false;
    private boolean decoderInicializado = false;
    private boolean pacoteInicial = false;

    //private long startMs;
    // If this stream is set, use it to trace packet arrival data
    //private OutputStream traceOutputStream = null;

    public RtpMediaDecoder(SurfaceView surfaceView) {

        log.info("RtpMediaDecoder started with params (" + DEBUGGING + "," + bufferType + "," + useNio + ","
                + receiveBufferSize + ")");

        this.surfaceView = surfaceView;
        surfaceView.getHolder().addCallback(this);
        //this.baos = new ByteArrayOutputStream();
    }

    /**
     * Defines the output stream where to trace packet's data while they arrive to the decoder
     *
     * @param outputStream stream where to dump data
     */
    /*public void setTraceOutputStream(OutputStream outputStream) {
    traceOutputStream = outputStream;
    }*/

    /**
     * Stops the underlying RTP session and properly releases the Android API decoder
     */
    public void release() {
        if (decoder != null) {
            try {
                decoder.stop();
            } catch (Exception e) {
                log.error("Encountered error while trying to stop decoder", e);
            }
            decoder.release();
            decoder = null;
        }
    }

    /**
     * Resizes surface view to 640x480
     *
     * @param holder
     */
    @Override
    public void surfaceCreated(SurfaceHolder holder) {
        /*android.view.ViewGroup.LayoutParams layoutParams = surfaceView.getLayoutParams();
        layoutParams.width = SURFACE_WIDTH; // required width
        layoutParams.height = SURFACE_HEIGHT; // required height
        surfaceView.setLayoutParams(layoutParams);*/

    }

    /**
     * Starts playing video when surface view is ready
     *
     * @param holder
     * @param format
     * @param width
     * @param height
     */
    @Override
    public void surfaceChanged(SurfaceHolder holder, int format, int width, int height) {
        if (playerThread == null) {
            playerThread = new PlayerThread(holder.getSurface());
            playerThread.setName("Received Video Thread");
            playerThread.start();
        }
    }

    @Override
    public void surfaceDestroyed(SurfaceHolder holder) {

    }

    public String getResolution() {
        return SURFACE_WIDTH + "x" + SURFACE_HEIGHT;
    }

    public String getTransportProtocol() {
        return TRANSPORT_PROTOCOL;
    }

    public String getVideoCodec() {
        return VIDEO_CODEC;
    }

    public int getDataStreamingPort() {
        return DATA_STREAMING_PORT;
    }

    public String getBufferType() {
        return bufferType;
    }

    /**
     * Creates the Android API decoder, configures it and starts it.
     */
    private class PlayerThread extends Thread {
        private Surface surface;

        /**
         * Thread constructor.
         *
         * @param surface where video will be played
         */
        public PlayerThread(Surface surface) {
            this.surface = surface;
        }

        @Override
        public void run() {
            // Wait a little bit to make sure the RtpClientThread had the opportunity to start
            // and create the rtpMediaExtractor

            MediaFormat mediaFormat = getMediaFormat();
            String mime = mediaFormat.getString(MediaFormat.KEY_MIME);
            if (mime.startsWith("video/")) {
                decoder = MediaCodec.createDecoderByType(mime);
                decoder.configure(mediaFormat, surface, null, 0);
            }

            if (decoder == null) {
                log.info("Can't find video info!");
                return;
            }

            decoder.start();

            inputBuffers = decoder.getInputBuffers();
            outputBuffers = decoder.getOutputBuffers();

            try {
                sleep(500);
            } catch (InterruptedException e) {
            }

            //startMs = System.currentTimeMillis();
        }

        /**
         * Decodes a frame
         *
         * @param decodeBuffer
         * @throws Exception
         */
        public void decodeFrame(ByteBuffer inputBuf, long timestamp, int offset, int inputBufIndex) {

            try {

                /*int inputBufIndex = decoder.dequeueInputBuffer(-1);
                 ByteBuffer inputBuf = inputBuffers[inputBufIndex];
                 inputBuf.clear();
                 inputBuf.put(packet);*/
                inputBuf.flip();

                // Queue the sample to be decoded
                decoder.queueInputBuffer(inputBufIndex, offset, inputBuf.remaining(), timestamp, 0);

                // Read the decoded output            
                int outIndex = decoder.dequeueOutputBuffer(info, 10000);
                switch (outIndex) {
                case MediaCodec.INFO_OUTPUT_BUFFERS_CHANGED:
                    if (DEBUGGING) {
                        log.info("The output buffers have changed.");
                    }
                    //outputBuffers = decoder.getOutputBuffers();
                    break;
                case MediaCodec.INFO_OUTPUT_FORMAT_CHANGED:
                    if (DEBUGGING) {
                        log.info("New format " + decoder.getOutputFormat());
                    }
                    break;
                case MediaCodec.INFO_TRY_AGAIN_LATER:
                    if (DEBUGGING) {
                        log.info("Call to dequeueOutputBuffer timed out.");
                    }
                    break;
                default:
                    if (DEBUGGING) {
                        //ByteBuffer buffer = outputBuffers[outIndex];
                        //log.info("We can't use this buffer but render it due to the API limit, " + buffer);
                    }

                    // return buffer to the codec
                    decoder.releaseOutputBuffer(outIndex, true);
                    break;
                }

            } catch (IllegalStateException e) {
                android.util.Log.e("VOICER", "Pacote invlido", e);
            }

            // All decoded frames have been rendered, we can stop playing now
            if (((info.flags & MediaCodec.BUFFER_FLAG_END_OF_STREAM) != 0) && DEBUGGING) {
                log.info("All decoded frames have been rendered");
            }
        }
    }

    /**
     * 
     * @return
     */
    public MediaFormat getMediaFormat() {
        String mimeType = "video/avc";

        MediaFormat format = MediaFormat.createVideoFormat(mimeType, SURFACE_WIDTH, SURFACE_HEIGHT);

        // from avconv, when streaming sample.h264.mp4 from disk
        /*byte[] header_sps = {0, 0, 0, 1, // header
            0x67, 0x64, (byte) 0x00, 0x1e, (byte) 0xac, (byte) 0xd9, 0x40, (byte) 0xa0, 0x3d,
            (byte) 0xa1, 0x00, 0x00, (byte) 0x03, 0x00, 0x01, 0x00, 0x00, 0x03, 0x00, 0x3C, 0x0F, 0x16, 0x2D, (byte) 0x96}; // sps
        byte[] header_pps = {0, 0, 0, 1, // header
            0x68, (byte) 0xeb, (byte) 0xec, (byte) 0xb2, 0x2C}; // pps
         */

        //byte[] header_sps = {0, 0, 0, 1, 103, 100, 0, 41, -84, 27, 26, -64, -96, 61, -112}; // sps
        //byte[] header_sps = {0, 0, 0, 1, 103, 66, 0, 30, -90, -128, -96, 61, -112}; //-samsung
        byte[] header_sps = { 0, 0, 0, 1, 103, 66, -128, 30, -85, 64, 80, 30, -48, -128, 0, 0, 3, 0, -128, 0, 0, 30,
                112, 32, 0, 122, 18, 0, 15, 66, 86, -79, -80, 16 }; //-asus
        //byte[] header_pps = {0, 0, 0, 1, 104, -22, 67, -53}; // pps
        //byte[] header_pps = {0, 0, 0, 1, 104, -50, 56, -128}; // -samsung
        byte[] header_pps = { 0, 0, 0, 1, 104, -50, 60, -128 }; // -asus

        /*format.setByteBuffer("csd-0", ByteBuffer.wrap(header_sps));
        format.setByteBuffer("csd-1", ByteBuffer.wrap(header_pps));*/

        format.setInteger(MediaFormat.KEY_MAX_INPUT_SIZE, SURFACE_WIDTH * SURFACE_HEIGHT);
        format.setInteger(MediaFormat.KEY_DURATION, 12600000);

        /*format.setInteger(MediaFormat.KEY_COLOR_FORMAT,
            MediaCodecInfo.CodecCapabilities.COLOR_FormatSurface);
        format.setInteger(MediaFormat.KEY_BIT_RATE, 1000000);
        format.setInteger(MediaFormat.KEY_FRAME_RATE, FRAME_RATE);
        format.setInteger(MediaFormat.KEY_I_FRAME_INTERVAL, IFRAME_INTERVAL);*/

        return format;
    }

    @Override
    public void processDatagramPacket(DatagramPacket pct) {
        if (playerThread == null || decoder == null)
            return;

        ChannelBuffer buffer = ChannelBuffers.wrappedBuffer(pct.getData(), 0, pct.getLength());
        DataPacket dp = DataPacket.decode(buffer);
        dp.setTimestamp(dp.getTimestamp() * 1000L / 90L);

        if (lastSequenceNumberIsValid && (lastSequenceNumber + 1) != dp.getSequenceNumber()) {
            lastSequenceNumber = dp.getSequenceNumber();
            return; // droppack
        }

        H264Packet h264Packet = new H264Packet(dp);

        if (!decoderInicializado && !h264Packet.h264NalType.equals(NalType.STAPA))
            return;

        switch (h264Packet.h264NalType) {
        case FULL:
            inputBufIndex = decoder.dequeueInputBuffer(-1);
            inputBuf = inputBuffers[inputBufIndex];
            inputBuf.clear();
            inputBuf.put(dp.getDataAsArray());

            playerThread.decodeFrame(inputBuf, dp.getTimestamp(), 1, inputBufIndex);
            break;
        case FUA:
            if (h264Packet.isStart()) {
                if (RtpMediaDecoder.DEBUGGING) {
                    log.info("FU-A start found. Starting new frame");
                }

                inputBufIndex = decoder.dequeueInputBuffer(-1);
                inputBuf = inputBuffers[inputBufIndex];
                inputBuf.clear();

                currentTimestamp = dp.getTimestamp();
            }

            // if we don't have a buffer here, it means that we skipped the start packet for this
            // NAL unit, so we can't do anything other than discard everything else
            if (currentTimestamp != 0) {

                // Did we miss packets in the middle of a frame transition?
                // In that case, I don't think there's much we can do other than flush our buffer
                // and discard everything until the next buffer
                if (dp.getTimestamp() != currentTimestamp) {
                    if (RtpMediaDecoder.DEBUGGING) {
                        log.warn("Non-consecutive timestamp found");
                    }
                } else {

                    inputBuf.put(dp.getDataAsArray(), 2, dp.getDataSize() - 2);

                }

                if (h264Packet.isEnd()) {
                    if (RtpMediaDecoder.DEBUGGING) {
                        log.info("FU-A end found. Sending frame!");
                    }

                    playerThread.decodeFrame(inputBuf, currentTimestamp, 0, inputBufIndex);

                    currentTimestamp = 0;

                }
            }
            break;
        case STAPA:

            decoderInicializado = true;

            inputBufIndex = decoder.dequeueInputBuffer(-1);
            inputBuf = inputBuffers[inputBufIndex];
            inputBuf.clear();

            int idx = VoicerHelper.indexOf(dp.getDataAsArray(), new byte[] { 0, 0, 0, 1 }, 4);

            inputBuf.put(dp.getDataAsArray(), 1, idx - 1);

            playerThread.decodeFrame(inputBuf, dp.getTimestamp(), 0, inputBufIndex);

            inputBufIndex = decoder.dequeueInputBuffer(-1);
            inputBuf = inputBuffers[inputBufIndex];
            inputBuf.clear();

            inputBuf.put(dp.getDataAsArray(), idx, dp.getDataSize() - idx);

            playerThread.decodeFrame(inputBuf, dp.getTimestamp(), 0, inputBufIndex);

            break;
        case UNKNOWN:
            break;
        default:
            break;
        }

        lastSequenceNumber = dp.getSequenceNumber();
        lastSequenceNumberIsValid = true;

        /*android.util.Log.d("VOICER", new String("<< Received: " + pct.getLength()));
        android.util.Log.d("VOICER", "<< HEX " + VoicerHelper.converteDadosBinariosParaStringHexa(dp.getDataAsArray()));
        android.util.Log.d("VOICER", "<< Sequence # " + dp.getSequenceNumber());*/

        /*String str = "";
        for (int i=0; i<pct.getLength(); i++)
           str+=pct.getData()[i] + ":";
        */
        //android.util.Log.d(VoicerHelper.TAG, "<< #" + dp.getSequenceNumber() + " length " + dp.getDataSize());

    }

    private enum NalType {
        FULL, FUA, STAPA, UNKNOWN
    }

    /**
     * H.264 Packet parsed following H.264 spec.
     */
    private class H264Packet {
        private final byte nalFBits;
        private final byte nalNriBits;
        private final byte nalType;
        private boolean fuStart = false;
        private boolean fuEnd = false;
        private byte fuNalType;
        private NalType h264NalType = NalType.UNKNOWN;

        /**
         * Creates a H.264 packet parsing its content
         *
         * @param packet
         */
        public H264Packet(DataPacket packet) {
            // Parsing the RTP Packet - http://www.ietf.org/rfc/rfc3984.txt section 5.3
            byte nalUnitOctet = packet.getData().getByte(0);
            nalFBits = (byte) (nalUnitOctet & 0x80);
            nalNriBits = (byte) (nalUnitOctet & 0x60);
            nalType = (byte) (nalUnitOctet & 0x1F);

            // If it's a single NAL packet then the entire payload is here
            if (nalType > 0 && nalType < 24) {
                h264NalType = NalType.FULL;
            } else if (nalType == 28) {
                h264NalType = NalType.FUA;

            } else if (nalType == 24) {
                h264NalType = NalType.STAPA;
            }

            byte fuHeader = packet.getData().getByte(1);
            fuStart = ((fuHeader & 0x80) != 0);
            fuEnd = ((fuHeader & 0x40) != 0);
            fuNalType = (byte) (fuHeader & 0x1F);
        }

        /**
         * Re-creates the H.264 NAL header for the FU-A header
         *
         * @return
         */
        public byte getNalTypeOctet() {
            // Excerpt from the spec:
            /* "The NAL unit type octet of the fragmented
               NAL unit is not included as such in the fragmentation unit payload,
               but rather the information of the NAL unit type octet of the
               fragmented NAL unit is conveyed in F and NRI fields of the FU
               indicator octet of the fragmentation unit and in the type field of
               the FU header"  */

            return (byte) (fuNalType | nalFBits | nalNriBits);
        }

        /**
         * Indicates whether this packet is the start of a frame.
         *
         * @return
         */
        public boolean isStart() {
            return fuStart;
        }

        /**
         * Indicates whether this packet is the end of a frame.
         *
         * @return
         */
        public boolean isEnd() {
            return fuEnd;
        }

        /**
         * Returns NAL type byte.
         *
         * @return
         */
        public byte getNalType() {
            return nalType;
        }
    }

}