summaryrefslogtreecommitdiff
path: root/sdl_android/src/main/java/com/smartdevicelink/encoder/SdlEncoder.java
blob: bd1e0c364650d1c5f92dd396333ff5f928ae5bdc (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
package com.smartdevicelink.encoder;

import java.io.IOException;
import java.io.PipedOutputStream;
import java.nio.ByteBuffer;

import android.annotation.TargetApi;
import android.media.MediaCodec;
import android.media.MediaCodecInfo;
import android.media.MediaFormat;
import android.os.Build;
import android.util.Log;
import android.view.Surface;

import com.smartdevicelink.proxy.interfaces.IVideoStreamListener;

@TargetApi(Build.VERSION_CODES.JELLY_BEAN_MR2)
public class SdlEncoder {

	private static final String TAG = "SdlEncoder";

	// parameters for the encoder
	private static final String _MIME_TYPE = "video/avc"; // H.264/AVC video
	// private static final String MIME_TYPE = "video/mp4v-es"; //MPEG4 video
	private int frameRate = 30;
	private int frameInterval = 5;
	private int frameWidth = 800;
	private int frameHeight = 480;
	private int bitrate = 6000000;

	// encoder state
	private MediaCodec mEncoder;
	private PipedOutputStream mOutputStream;
	private IVideoStreamListener mOutputListener;
	
	// allocate one of these up front so we don't need to do it every time
	private MediaCodec.BufferInfo mBufferInfo;

	// Codec-specific data (SPS and PPS)
	private byte[] mH264CodecSpecificData = null;

	public SdlEncoder () {
	}
	public void setFrameRate(int iVal){
		frameRate = iVal;
	}
	public void setFrameInterval(int iVal){
		frameInterval = iVal;
	}
	public void setFrameWidth(int iVal){
		frameWidth = iVal;	
	}
	public void setFrameHeight(int iVal){
		frameHeight = iVal;
	}
	public void setBitrate(int iVal){
		bitrate = iVal;	
	}
	public void setOutputStream(PipedOutputStream mStream){
		mOutputStream = mStream;
	}
	public void setOutputListener(IVideoStreamListener listener) {
		mOutputListener = listener;
	}
	public Surface prepareEncoder () {

		mBufferInfo = new MediaCodec.BufferInfo();

		MediaFormat format = MediaFormat.createVideoFormat(_MIME_TYPE, frameWidth,
				frameHeight);

		// Set some properties. Failing to specify some of these can cause the
		// MediaCodec
		// configure() call to throw an unhelpful exception.
		format.setInteger(MediaFormat.KEY_COLOR_FORMAT,
				MediaCodecInfo.CodecCapabilities.COLOR_FormatSurface);
		format.setInteger(MediaFormat.KEY_BIT_RATE, bitrate);
		format.setInteger(MediaFormat.KEY_FRAME_RATE, frameRate);
		format.setInteger(MediaFormat.KEY_I_FRAME_INTERVAL, frameInterval);

		// Create a MediaCodec encoder, and configure it with our format. Get a
		// Surface
		// we can use for input and wrap it with a class that handles the EGL
		// work.
		//
		// If you want to have two EGL contexts -- one for display, one for
		// recording --
		// you will likely want to defer instantiation of CodecInputSurface
		// until after the
		// "display" EGL context is created, then modify the eglCreateContext
		// call to
		// take eglGetCurrentContext() as the share_context argument.
		try {
			mEncoder = MediaCodec.createEncoderByType(_MIME_TYPE);
		} catch (Exception e) {e.printStackTrace();}

		if(mEncoder != null) {
		   mEncoder.configure(format, null, null, MediaCodec.CONFIGURE_FLAG_ENCODE);
           return mEncoder.createInputSurface();
		} else {
			return null;
		}
	}
	
	public void startEncoder () {
		if(mEncoder != null) {
		  mEncoder.start();
		}
	}
	
	/**
	 * Releases encoder resources.
	 */
	public void releaseEncoder() {
		if (mEncoder != null) {
			mEncoder.stop();
			mEncoder.release();
			mEncoder = null;
		}
		if (mOutputStream != null) {
			try {
				mOutputStream.close();
			} catch (IOException e) {
				e.printStackTrace();
			}
			mOutputStream = null;
		}
		mH264CodecSpecificData = null;
	}

	/**
	 * Extracts all pending data from the encoder
	 * <p>
	 * If endOfStream is not set, this returns when there is no more data to
	 * drain. If it is set, we send EOS to the encoder, and then iterate until
	 * we see EOS on the output. Calling this with endOfStream set should be
	 * done once, right before stopping the muxer.
	 */
	public void drainEncoder(boolean endOfStream) {
		final int TIMEOUT_USEC = 10000;

		if(mEncoder == null || (mOutputStream == null && mOutputListener == null)) {
		   return;			
		}
		if (endOfStream) {
			  mEncoder.signalEndOfInputStream();
		}

		ByteBuffer[] encoderOutputBuffers = mEncoder.getOutputBuffers();
		while (true) {
			int encoderStatus = mEncoder.dequeueOutputBuffer(mBufferInfo,
					TIMEOUT_USEC);
			if (encoderStatus == MediaCodec.INFO_TRY_AGAIN_LATER) {
				// no output available yet
				if (!endOfStream) {
					break; // out of while
				}
			} else if (encoderStatus == MediaCodec.INFO_OUTPUT_BUFFERS_CHANGED) {
				// not expected for an encoder
				encoderOutputBuffers = mEncoder.getOutputBuffers();
			} else if (encoderStatus == MediaCodec.INFO_OUTPUT_FORMAT_CHANGED) {
				if (mH264CodecSpecificData == null) {
					MediaFormat format = mEncoder.getOutputFormat();
					mH264CodecSpecificData = EncoderUtils.getCodecSpecificData(format);
				} else {
					Log.w(TAG, "Output format change notified more than once, ignoring.");
				}
			} else if (encoderStatus < 0) {
			} else {
				if ((mBufferInfo.flags & MediaCodec.BUFFER_FLAG_CODEC_CONFIG) != 0) {
					// If we already retrieve codec specific data via OUTPUT_FORMAT_CHANGED event,
					// we do not need this data.
					if (mH264CodecSpecificData != null) {
						mBufferInfo.size = 0;
					} else {
						Log.i(TAG, "H264 codec specific data not retrieved yet.");
					}
				}

				if (mBufferInfo.size != 0) {
					ByteBuffer encoderOutputBuffer = encoderOutputBuffers[encoderStatus];
					byte[] dataToWrite = null;
					int dataOffset = 0;

					// append SPS and PPS in front of every IDR NAL unit
					if ((mBufferInfo.flags & MediaCodec.BUFFER_FLAG_KEY_FRAME) != 0
							&& mH264CodecSpecificData != null) {
						dataToWrite = new byte[mH264CodecSpecificData.length + mBufferInfo.size];
						System.arraycopy(mH264CodecSpecificData, 0,
								dataToWrite, 0, mH264CodecSpecificData.length);
						dataOffset = mH264CodecSpecificData.length;
					} else {
						dataToWrite = new byte[mBufferInfo.size];
					}

					try {
						encoderOutputBuffer.position(mBufferInfo.offset);
						encoderOutputBuffer.limit(mBufferInfo.offset + mBufferInfo.size);

						encoderOutputBuffer.get(dataToWrite, dataOffset, mBufferInfo.size);

						if (mOutputStream != null) {
							mOutputStream.write(dataToWrite, 0, mBufferInfo.size);
						} else if (mOutputListener != null) {
							mOutputListener.sendFrame(
									dataToWrite, 0, dataToWrite.length, mBufferInfo.presentationTimeUs);
						}
					} catch (Exception e) {}
				}

				mEncoder.releaseOutputBuffer(encoderStatus, false);

				if ((mBufferInfo.flags & MediaCodec.BUFFER_FLAG_END_OF_STREAM) != 0) {
					break; // out of while
				}
			}
		}
	}
}