Recherche avancée

Médias (0)

Mot : - Tags -/optimisation

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (79)

  • Le profil des utilisateurs

    12 avril 2011, par

    Chaque utilisateur dispose d’une page de profil lui permettant de modifier ses informations personnelle. Dans le menu de haut de page par défaut, un élément de menu est automatiquement créé à l’initialisation de MediaSPIP, visible uniquement si le visiteur est identifié sur le site.
    L’utilisateur a accès à la modification de profil depuis sa page auteur, un lien dans la navigation "Modifier votre profil" est (...)

  • Configurer la prise en compte des langues

    15 novembre 2010, par

    Accéder à la configuration et ajouter des langues prises en compte
    Afin de configurer la prise en compte de nouvelles langues, il est nécessaire de se rendre dans la partie "Administrer" du site.
    De là, dans le menu de navigation, vous pouvez accéder à une partie "Gestion des langues" permettant d’activer la prise en compte de nouvelles langues.
    Chaque nouvelle langue ajoutée reste désactivable tant qu’aucun objet n’est créé dans cette langue. Dans ce cas, elle devient grisée dans la configuration et (...)

  • XMP PHP

    13 mai 2011, par

    Dixit Wikipedia, XMP signifie :
    Extensible Metadata Platform ou XMP est un format de métadonnées basé sur XML utilisé dans les applications PDF, de photographie et de graphisme. Il a été lancé par Adobe Systems en avril 2001 en étant intégré à la version 5.0 d’Adobe Acrobat.
    Étant basé sur XML, il gère un ensemble de tags dynamiques pour l’utilisation dans le cadre du Web sémantique.
    XMP permet d’enregistrer sous forme d’un document XML des informations relatives à un fichier : titre, auteur, historique (...)

Sur d’autres sites (5689)

  • FFMPEG C api h.264 encoding / MPEG2 ts streaming problems

    3 mars 2015, par ccoral

    Class prototype is as follows :

    #ifndef _FULL_MOTION_VIDEO_STREAM_H_
    #define _FULL_MOTION_VIDEO_STREAM_H_

    #include <memory>
    #include <string>

    #ifndef INT64_C
    # define INT64_C(c) (c ## LL)
    # define UINT64_C(c) (c ## ULL)
    #endif

    extern "C"
    {
       #include "libavutil/opt.h"
       #include "libavcodec/avcodec.h"
       #include "libavutil/channel_layout.h"
       #include "libavutil/common.h"
       #include "libavutil/imgutils.h"
       #include "libavutil/mathematics.h"
       #include "libavutil/samplefmt.h"
       #include "libavformat/avformat.h"

       #include <libavutil></libavutil>timestamp.h>
       #include <libswscale></libswscale>swscale.h>
       #include <libswresample></libswresample>swresample.h>
    }

    class FMVStream
    {
       public:
           struct OutputStream
           {
               OutputStream() :
               st(0),
               next_pts(0),
               samples_count(0),
               frame(0),
               tmpFrame(0),
               sws_ctx(0)
               {
               }

               AVStream *st;

               /* pts of the next frame that will be generated */
               int64_t next_pts;
               int samples_count;

               AVFrame *frame;
               AVFrame *tmpFrame;

               struct SwsContext *sws_ctx;
           };

           ///
           /// Constructor
           ///
           FMVStream();

           ///
           /// Destructor
           ///
           ~FMVStream();

           ///
           /// Frame encoder helper function
           ///
           /// Encodes a raw RGB frame into the transport stream
           ///
           int EncodeFrame(uint8_t* frame);

           ///
           /// Frame width setter
           ///
           void setFrameWidth(int width);

           ///
           /// Frame width getter
           ///
           int getFrameWidth() const;

           ///
           /// Frame height setter
           ///
           void setFrameHeight(int height);

           ///
           /// Frame height getter
           ///
           int getFrameHeight() const;

           ///
           /// Stream address setter
           ///
           void setStreamAddress(const std::string&amp; address);

           ///
           /// Stream address getter
           ///
           std::string getStreamAddress() const;

       private:

           ///
           /// Video Stream creation
           ///
           AVStream* initVideoStream(AVFormatContext* oc);

           ///
           /// Raw frame transcoder
           ///
           /// This will convert the raw RGB frame to a raw YUV frame necessary for h.264 encoding
           ///
           void CopyFrameData(uint8_t* src_frame);

           ///
           /// Video frame allocator
           ///
           AVFrame* AllocPicture(PixelFormat pix_fmt, int width, int height);

           ///
           /// Debug print helper function
           ///
           void print_sdp(AVFormatContext **avc, int n);

           ///
           /// Write the frame to the stream
           ///
           int write_frame(AVFormatContext *fmt_ctx, const AVRational *time_base, AVStream *st, AVPacket *pkt);

           ///
           /// initialize the frame data
           ///
           void initFrame();

           // formatting data needed for output streaming and the output container (MPEG 2 TS)
           AVOutputFormat* format;
           AVFormatContext* format_ctx;

           // structure container for our video stream
           OutputStream stream;

           AVIOContext* io_ctx;

           std::string streamFilename;

           int frameWidth;
           int frameHeight;
    };

    #endif
    </string></memory>

    This block starts the class declaration.

    #include "FullMotionVideoStream.h"

    #include <stdexcept>
    #include <iostream>

    FMVStream::FMVStream()
       : format(0),
       format_ctx(0),
       stream(),
       io_ctx(0),
       streamFilename("test.mpeg"),
       frameWidth(640),
       frameHeight(480)
    {
       // Register all formats and codecs
       av_register_all();
       avcodec_register_all();

       // Init networking
       avformat_network_init();

       // Find format
       this->format = av_guess_format("mpegts", NULL, NULL);

       // allocate the AVFormatContext
       this->format_ctx = avformat_alloc_context();

       if (!this->format_ctx)
       {
           throw std::runtime_error("avformat_alloc_context failed");
       }

       this->format_ctx->oformat = this->format;
       //sprintf_s(this->format_ctx->filename, sizeof(this->format_ctx->filename), "%s", this->streamFilename.c_str());

       this->stream.st = initVideoStream(this->format_ctx);

       this->initFrame();

       // Allocate AVIOContext
       int ret = avio_open(&amp;this->io_ctx, this->streamFilename.c_str(), AVIO_FLAG_WRITE);

       if (ret != 0)
       {
           throw std::runtime_error("avio_open failed");
       }

       this->format_ctx->pb = this->io_ctx;

       // Print some debug info about the format
       av_dump_format(this->format_ctx, 0, NULL, 1);

       // Begin the output by writing the container header
       avformat_write_header(this->format_ctx, NULL);

       AVFormatContext* ac[] = { this->format_ctx };
       print_sdp(ac, 1);
    }

    FMVStream::~FMVStream()
    {
       av_write_trailer(this->format_ctx);
       avcodec_close(this->stream.st->codec);

       avio_close(io_ctx);

       avformat_free_context(this->format_ctx);

       av_frame_free(&amp;this->stream.frame);
       av_free(this->format);
    }

    AVFrame* FMVStream::AllocPicture(PixelFormat pix_fmt, int width, int height)
    {
       // Allocate a frame
       AVFrame* frame = av_frame_alloc();

       if (frame == nullptr)
       {
           throw std::runtime_error("avcodec_alloc_frame failed");
       }

       if (av_image_alloc(frame->data, frame->linesize, width, height, pix_fmt, 1) &lt; 0)
       {
           throw std::runtime_error("av_image_alloc failed");
       }

       frame->width = width;
       frame->height = height;
       frame->format = pix_fmt;

       return frame;
    }

    void FMVStream::print_sdp(AVFormatContext **avc, int n)
    {
       char sdp[2048];
       av_sdp_create(avc, n, sdp, sizeof(sdp));
       printf("SDP:\n%s\n", sdp);
       fflush(stdout);
    }

    AVStream* FMVStream::initVideoStream(AVFormatContext *oc)
    {
       AVStream* st = avformat_new_stream(oc, NULL);

       if (st == nullptr)
       {
           std::runtime_error("Could not alloc stream");
       }

       AVCodec* codec = avcodec_find_encoder(AV_CODEC_ID_H264);

       if (codec == nullptr)
       {
           throw std::runtime_error("couldn't find mpeg2 encoder");
       }

       st->codec = avcodec_alloc_context3(codec);

       st->codec->codec_id = AV_CODEC_ID_H264;
       st->codec->codec_type = AVMEDIA_TYPE_VIDEO;
       st->codec->bit_rate = 400000;

       st->codec->width = this->frameWidth;
       st->codec->height = this->frameHeight;

       st->time_base.num = 1;
       st->time_base.den = 30;

       st->codec->framerate.num = 1;
       st->codec->framerate.den = 30;

       st->codec->max_b_frames = 2;
       st->codec->gop_size = 12;
       st->codec->pix_fmt = PIX_FMT_YUV420P;

       st->id = oc->nb_streams - 1;

       if (oc->oformat->flags &amp; AVFMT_GLOBALHEADER)
       {
           st->codec->flags |= CODEC_FLAG_GLOBAL_HEADER;
       }

       // option setup for the codec
       av_opt_set(st->codec->priv_data, "profile", "baseline", AV_OPT_SEARCH_CHILDREN);

       if (avcodec_open2(st->codec, codec, NULL) &lt; 0)
       {
           throw std::runtime_error("avcodec_open failed");
       }

       return st;
    }

    void FMVStream::initFrame()
    {
       // Allocate a tmp frame for converting our raw RGB data to YUV for encoding
       this->stream.tmpFrame = this->AllocPicture(PIX_FMT_RGB24, this->frameWidth, this->frameHeight);

       // Allocate a main frame
       this->stream.frame = this->AllocPicture(PIX_FMT_YUV420P, this->frameWidth, this->frameHeight);
    }
    </iostream></stdexcept>

    This block is attempting to convert from the raw RGB to our needed YUV format for h.264 encoding.

    void FMVStream::CopyFrameData(uint8_t* data)
    {
       // fill image with our raw RGB data
       //avpicture_alloc((AVPicture*)this->stream.tmpFrame, PIX_FMT_RGB24, this->stream.st->codec->width, this->stream.st->codec->height);

       int numBytes = avpicture_get_size(PIX_FMT_RGB24, this->stream.st->codec->width, this->stream.st->codec->height);

       uint8_t* buffer = (uint8_t*) av_malloc(numBytes * sizeof(uint8_t));

       avpicture_fill((AVPicture*)this->stream.tmpFrame, buffer, PIX_FMT_RGB24, this->stream.st->codec->width, this->stream.st->codec->height);

       for (int y = 0; y &lt; this->stream.st->codec->height; y++)
       {
           for (int x = 0; x &lt; this->stream.st->codec->width; x++)
           {
               int offset = 3 * (x + y * this->stream.st->codec->width);
               this->stream.tmpFrame->data[0][offset + 0] = data[x + y * this->stream.st->codec->width]; // R
               this->stream.tmpFrame->data[0][offset + 1] = data[x + y * this->stream.st->codec->width + 1]; // G
               this->stream.tmpFrame->data[0][offset + 2] = data[x + y * this->stream.st->codec->width + 2]; // B
           }
       }

       // convert the RGB frame to a YUV frame using the sws Context
       this->stream.sws_ctx = sws_getContext(this->stream.st->codec->width, this->stream.st->codec->height, PIX_FMT_RGB32, this->stream.st->codec->width, this->stream.st->codec->height, PIX_FMT_YUV420P, SWS_FAST_BILINEAR, NULL, NULL, NULL);

       // use the scale function to transcode this raw frame to the correct type
       sws_scale(this->stream.sws_ctx, this->stream.tmpFrame->data, this->stream.tmpFrame->linesize, 0, this->stream.st->codec->height, this->stream.frame->data, this->stream.frame->linesize);
    }

    This is the block that encodes the raw data to h.264, and then send it out the Mpeg2 ts. I believe the problem lies within this block. I can put a break point in my write frame block and see that frames are being written, however, opening the resulting file in VLC results in a blank video. The file is approx 2Mb.

    int FMVStream::EncodeFrame(uint8_t* data)
    {
       AVCodecContext* c = this->stream.st->codec;

       AVRational one;
       one.den = one.num = 1;

       // check to see if we want to keep writing frames we can probably change this to a toggle switch
       if (av_compare_ts(this->stream.next_pts, this->stream.st->codec->time_base, 10, one) >= 0)
       {
           this->stream.frame = nullptr;
       }
       else
       {
           // Convert and load the frame data into the AVFrame struct
           CopyFrameData(data);
       }

       // setup the timestamp stepping
       AVPacket pkt = { 0 };
       av_init_packet(&amp;pkt);
       this->stream.frame->pts = (int64_t)((1.0 / this->stream.st->codec->framerate.den) * 90000.0 * this->stream.next_pts++);

       int gotPacket, out_size, ret;

       out_size = avcodec_encode_video2(c, &amp;pkt, this->stream.frame, &amp;gotPacket);


       if (gotPacket == 1)
       {
           ret = write_frame(this->format_ctx, &amp;c->time_base, this->stream.st, &amp;pkt);
       }
       else
       {
           ret = 0;
       }

       if (ret &lt; 0)
       {
           std::cerr &lt;&lt; "Error writing video frame" &lt;&lt; std::endl;
       }

       av_free_packet(&amp;pkt);

       return ((this->stream.frame != nullptr) || gotPacket) ? 0 : 1;
    }

    int FMVStream::write_frame(AVFormatContext *fmt_ctx, const AVRational *time_base, AVStream *st, AVPacket *pkt)
    {
       /* rescale output packet timestamp values from codec to stream timebase */
       av_packet_rescale_ts(pkt, *time_base, st->time_base);
       pkt->stream_index = st->index;

       return av_interleaved_write_frame(fmt_ctx, pkt);
    }

    void FMVStream::setFrameWidth(const int width)
    {
       this->frameWidth = width;
    }

    int FMVStream::getFrameWidth() const
    {
       return this->frameWidth;
    }

    void FMVStream::setFrameHeight(const int height)
    {
       this->frameHeight = height;
    }

    int FMVStream::getFrameHeight() const
    {
       return this->frameHeight;
    }

    void FMVStream::setStreamAddress(const std::string&amp; address)
    {
       this->streamFilename = address;
    }

    std::string FMVStream::getStreamAddress() const
    {
       return this->streamFilename;
    }

    Here is the Main function.

    #include "FullMotionVideoStream.h"

    #include <iostream>
    #include <thread>
    #include <chrono>

    int main(int argc, char** argv)
    {
       FMVStream* fmv = new FMVStream;

       fmv->setFrameWidth(640);
       fmv->setFrameHeight(480);

       std::cout &lt;&lt; "Streaming Address: " &lt;&lt; fmv->getStreamAddress() &lt;&lt; std::endl;

       // create our alternating frame of black and white to test the streaming functionality
       uint8_t white[640 * 480 * sizeof(uint8_t) * 3];
       uint8_t black[640 * 480 * sizeof(uint8_t) * 3];

       std::memset(white, 255, 640 * 480 * sizeof(uint8_t) * 3);
       std::memset(black, 0, 640 * 480 * sizeof(uint8_t)* 3);

       for (auto i = 0; i &lt; 100; i++)
       {
           auto ret = fmv->EncodeFrame(white);

           if (ret != 0)
           {
               std::cerr &lt;&lt; "There was a problem encoding the frame: " &lt;&lt; i &lt;&lt; std::endl;
           }

           std::this_thread::sleep_for(std::chrono::milliseconds(10));
       }

       for (auto i = 0; i &lt; 100; i++)
       {
           auto ret = fmv->EncodeFrame(black);

           if (ret != 0)
           {
               std::cerr &lt;&lt; "There was a problem encoding the frame: " &lt;&lt; i &lt;&lt; std::endl;
           }

           std::this_thread::sleep_for(std::chrono::milliseconds(10));
       }

       delete fmv;
    }
    </chrono></thread></iostream>

    Here is the resultant output via the console / my print SDP function.

    [libx264 @ 000000ac95f58440] using cpu capabilities: MMX2 SSE2Fast SSSE3 SSE4.2
    AVX FMA3 AVX2 LZCNT BMI2
    [libx264 @ 000000ac95f58440] profile Constrained Baseline, level 3.0
    Output #0, mpegts, to '(null)':
       Stream #0:0: Video: h264 (libx264), yuv420p, 640x480, q=-1--1, 400 kb/s, 30
    tbn
    SDP:
    v=0
    o=- 0 0 IN IP4 127.0.0.1
    s=No Name
    t=0 0
    a=tool:libavformat 56.23.104
    m=video 0 RTP/AVP 96
    b=AS:400
    a=rtpmap:96 H264/90000
    a=fmtp:96 packetization-mode=1
    a=control:streamid=0

    Streaming Address: test.mpeg
    [libx264 @ 000000ac95f58440] frame I:45    Avg QP: 0.51  size:  1315
    [libx264 @ 000000ac95f58440] frame P:136   Avg QP: 0.29  size:   182
    [libx264 @ 000000ac95f58440] mb I  I16..4: 99.7%  0.0%  0.3%
    [libx264 @ 000000ac95f58440] mb P  I16..4:  0.1%  0.0%  0.1%  P16..4:  0.1%  0.0
    %  0.0%  0.0%  0.0%    skip:99.7%
    [libx264 @ 000000ac95f58440] final ratefactor: -68.99
    [libx264 @ 000000ac95f58440] coded y,uvDC,uvAC intra: 0.5% 0.5% 0.5% inter: 0.0%
    0.1% 0.1%
    [libx264 @ 000000ac95f58440] i16 v,h,dc,p: 96%  0%  3%  0%
    [libx264 @ 000000ac95f58440] i4 v,h,dc,ddl,ddr,vr,hd,vl,hu:  1% 10% 85%  0%  3%
    0%  1%  0%  0%
    [libx264 @ 000000ac95f58440] i8c dc,h,v,p: 100%  0%  0%  0%
    [libx264 @ 000000ac95f58440] ref P L0: 46.8% 25.2% 28.0%
    [libx264 @ 000000ac95f58440] kb/s:0.03

    I know there are probably many issues with this program, I am very new with FFMPEG and multimedia programming in general. Ive used many pieces of code found through searching google/ stack overflow to get to this point as is. The file has a good size but comes up as length 0.04 tells me that my time stamping must be broken between the frames / pkts, but I am unsure on how to fix this issue.

    I tried inspecting the file with ffmpeg.exe using ffmpeg -i and outputting to a regular TS. It seems my code works more then I originally intended however, I am simply trying to output a bunch of all white frames.

    ffmpeg -i test.mpeg test.ts
    ffmpeg version N-70125-g6c9537b Copyright (c) 2000-2015 the FFmpeg developers
     built with gcc 4.9.2 (GCC)
     configuration: --disable-static --enable-shared --enable-gpl --enable-version3
    --disable-w32threads --enable-avisynth --enable-bzlib --enable-fontconfig --ena
    ble-frei0r --enable-gnutls --enable-iconv --enable-libass --enable-libbluray --e
    nable-libbs2b --enable-libcaca --enable-libfreetype --enable-libgme --enable-lib
    gsm --enable-libilbc --enable-libmodplug --enable-libmp3lame --enable-libopencor
    e-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libopus --enabl
    e-librtmp --enable-libschroedinger --enable-libsoxr --enable-libspeex --enable-l
    ibtheora --enable-libtwolame --enable-libvidstab --enable-libvo-aacenc --enable-
    libvo-amrwbenc --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-l
    ibwebp --enable-libx264 --enable-libx265 --enable-libxavs --enable-libxvid --ena
    ble-lzma --enable-decklink --enable-zlib
     libavutil      54. 19.100 / 54. 19.100
     libavcodec     56. 26.100 / 56. 26.100
     libavformat    56. 23.104 / 56. 23.104
     libavdevice    56.  4.100 / 56.  4.100
     libavfilter     5. 11.101 /  5. 11.101
     libswscale      3.  1.101 /  3.  1.101
     libswresample   1.  1.100 /  1.  1.100
     libpostproc    53.  3.100 / 53.  3.100
    Input #0, mpegts, from 'test.mpeg':
     Duration: 00:00:00.04, start: 0.000000, bitrate: 24026 kb/s
     Program 1
       Metadata:
         service_name    : Service01
         service_provider: FFmpeg
       Stream #0:0[0x100]: Video: h264 (Constrained Baseline) ([27][0][0][0] / 0x00
    1B), yuv420p, 640x480, 25 fps, 25 tbr, 90k tbn, 50 tbc
    File 'test.ts' already exists. Overwrite ? [y/N] y
    Output #0, mpegts, to 'test.ts':
     Metadata:
       encoder         : Lavf56.23.104
       Stream #0:0: Video: mpeg2video, yuv420p, 640x480, q=2-31, 200 kb/s, 25 fps,
    90k tbn, 25 tbc
       Metadata:
         encoder         : Lavc56.26.100 mpeg2video
    Stream mapping:
     Stream #0:0 -> #0:0 (h264 (native) -> mpeg2video (native))
    Press [q] to stop, [?] for help
    frame=    3 fps=0.0 q=2.0 Lsize=       9kB time=00:00:00.08 bitrate= 883.6kbits/
    s dup=0 drop=178
    video:7kB audio:0kB subtitle:0kB other streams:0kB global headers:0kB muxing ove
    rhead: 22.450111%
  • Popcorn Hour revisited

    26 février 2015, par Mans — Hardware

    In a previous post, I recounted my adventures with the Popcorn Hour C-200 media player. Having not given the device much thought since then, I recently decided to dust it off and see what it was really made of. This is what I found. Components The system is based on … Continue reading

  • Muxing Android MediaCodec encoded H264 packets into RTMP

    31 décembre 2015, par Vadym

    I am coming from a thread Encoding H.264 from camera with Android MediaCodec. My setup is very similar. However, I attempt to write mux the encoded frames and with javacv and broadcast them via rtmp.

    RtmpClient.java

    ...
    private volatile BlockingQueue mFrameQueue = new LinkedBlockingQueue(MAXIMUM_VIDEO_FRAME_BACKLOG);
    ...
    private void startStream() throws FrameRecorder.Exception, IOException {
       if (TextUtils.isEmpty(mDestination)) {
           throw new IllegalStateException("Cannot start RtmpClient without destination");
       }

       if (mCamera == null) {
           throw new IllegalStateException("Cannot start RtmpClient without camera.");
       }

       Camera.Parameters cameraParams = mCamera.getParameters();

       mRecorder = new FFmpegFrameRecorder(
               mDestination,
               mVideoQuality.resX,
               mVideoQuality.resY,
               (mAudioQuality.channelType.equals(AudioQuality.CHANNEL_TYPE_STEREO) ? 2 : 1));

       mRecorder.setFormat("flv");

       mRecorder.setFrameRate(mVideoQuality.frameRate);
       mRecorder.setVideoBitrate(mVideoQuality.bitRate);
       mRecorder.setVideoCodec(avcodec.AV_CODEC_ID_H264);

       mRecorder.setSampleRate(mAudioQuality.samplingRate);
       mRecorder.setAudioBitrate(mAudioQuality.bitRate);
       mRecorder.setAudioCodec(avcodec.AV_CODEC_ID_AAC);

       mVideoStream = new VideoStream(mRecorder, mVideoQuality, mFrameQueue, mCamera);
       mAudioStream = new AudioStream(mRecorder, mAudioQuality);

       mRecorder.start();

       // Setup a bufferred preview callback
       setupCameraCallback(mCamera, mRtmpClient, DEFAULT_PREVIEW_CALLBACK_BUFFERS,
               mVideoQuality.resX * mVideoQuality.resY * ImageFormat.getBitsPerPixel(
                       cameraParams.getPreviewFormat())/8);

       try {
           mVideoStream.start();
           mAudioStream.start();
       }
       catch(Exception e) {
           e.printStackTrace();
           stopStream();
       }
    }
    ...
    @Override
    public void onPreviewFrame(byte[] data, Camera camera) {
       boolean frameQueued = false;

       if (mRecorder == null || data == null) {
           return;
       }

       frameQueued = mFrameQueue.offer(data);

       // return the buffer to be reused - done in videostream
       //camera.addCallbackBuffer(data);
    }
    ...

    VideoStream.java

    ...
    @Override
    public void run() {
       try {
           mMediaCodec = MediaCodec.createEncoderByType("video/avc");
           MediaFormat mediaFormat = MediaFormat.createVideoFormat("video/avc", mVideoQuality.resX, mVideoQuality.resY);
           mediaFormat.setInteger(MediaFormat.KEY_BIT_RATE, mVideoQuality.bitRate);
           mediaFormat.setInteger(MediaFormat.KEY_FRAME_RATE, mVideoQuality.frameRate);
           mediaFormat.setInteger(MediaFormat.KEY_COLOR_FORMAT, MediaCodecInfo.CodecCapabilities.COLOR_FormatYUV420SemiPlanar);
           mediaFormat.setInteger(MediaFormat.KEY_I_FRAME_INTERVAL, 1);
           mMediaCodec.configure(mediaFormat, null, null, MediaCodec.CONFIGURE_FLAG_ENCODE);
           mMediaCodec.start();
       }
       catch(IOException e) {
           e.printStackTrace();
       }

       long startTimestamp = System.currentTimeMillis();
       long frameTimestamp = 0;
       byte[] rawFrame = null;

       try {
           while (!Thread.interrupted()) {
               rawFrame = mFrameQueue.take();

               frameTimestamp = 1000 * (System.currentTimeMillis() - startTimestamp);

               encodeFrame(rawFrame, frameTimestamp);

               // return the buffer to be reused
               mCamera.addCallbackBuffer(rawFrame);
           }
       }
       catch (InterruptedException ignore) {
           // ignore interrup while waiting
       }

       // Clean up video stream allocations
       try {
           mMediaCodec.stop();
           mMediaCodec.release();
           mOutputStream.flush();
           mOutputStream.close();
       } catch (Exception e){
           e.printStackTrace();
       }
    }
    ...
    private void encodeFrame(byte[] input, long timestamp) {
       try {
           ByteBuffer[] inputBuffers = mMediaCodec.getInputBuffers();
           ByteBuffer[] outputBuffers = mMediaCodec.getOutputBuffers();

           int inputBufferIndex = mMediaCodec.dequeueInputBuffer(0);

           if (inputBufferIndex >= 0) {
               ByteBuffer inputBuffer = inputBuffers[inputBufferIndex];
               inputBuffer.clear();
               inputBuffer.put(input);
               mMediaCodec.queueInputBuffer(inputBufferIndex, 0, input.length, timestamp, 0);
           }

           MediaCodec.BufferInfo bufferInfo = new MediaCodec.BufferInfo();

           int outputBufferIndex = mMediaCodec.dequeueOutputBuffer(bufferInfo, 0);

           if (outputBufferIndex >= 0) {
               while (outputBufferIndex >= 0) {
                   ByteBuffer outputBuffer = outputBuffers[outputBufferIndex];

                   // Should this be a direct byte buffer?
                   byte[] outData = new byte[bufferInfo.size - bufferInfo.offset];
                   outputBuffer.get(outData);

                   mFrameRecorder.record(outData, bufferInfo.offset, outData.length, timestamp);

                   mMediaCodec.releaseOutputBuffer(outputBufferIndex, false);
                   outputBufferIndex = mMediaCodec.dequeueOutputBuffer(bufferInfo, 0);
               }
           }
           else if (outputBufferIndex == MediaCodec.INFO_OUTPUT_BUFFERS_CHANGED) {
               outputBuffers = mMediaCodec.getOutputBuffers();
           } else if (outputBufferIndex == MediaCodec.INFO_OUTPUT_FORMAT_CHANGED) {
               // ignore for now
           }
       } catch (Throwable t) {
           t.printStackTrace();
       }

    }
    ...

    FFmpegFrameRecorder.java

    ...
    // Hackish codec copy frame recording function
    public boolean record(byte[] encodedData, int offset, int length, long frameCount) throws Exception {
       int ret;

       if (encodedData == null) {
           return false;
       }

       av_init_packet(video_pkt);

       // this is why i wondered whether I should get outputbuffer data into direct byte buffer
       video_outbuf.put(encodedData, 0, encodedData.length);

       video_pkt.data(video_outbuf);
       video_pkt.size(video_outbuf_size);

       video_pkt.pts(frameCount);
       video_pkt.dts(frameCount);

       video_pkt.stream_index(video_st.index());

       synchronized (oc) {
           /* write the compressed frame in the media file */
           if (interleaved &amp;&amp; audio_st != null) {
               if ((ret = av_interleaved_write_frame(oc, video_pkt)) &lt; 0) {
                   throw new Exception("av_interleaved_write_frame() error " + ret + " while writing interleaved video frame.");
               }
           } else {
               if ((ret = av_write_frame(oc, video_pkt)) &lt; 0) {
                   throw new Exception("av_write_frame() error " + ret + " while writing video frame.");
               }
           }
       }
       return (video_pkt.flags() &amp; AV_PKT_FLAG_KEY) == 1;
    }
    ...

    When I try to stream the video and run ffprobe on it, I get the following output :

    ffprobe version 2.5.3 Copyright (c) 2007-2015 the FFmpeg developers
     built on Jan 19 2015 12:56:57 with gcc 4.1.2 (GCC) 20080704 (Red Hat 4.1.2-55)
     configuration: --prefix=/usr --bindir=/usr/bin --datadir=/usr/share/ffmpeg --incdir=/usr/include/ffmpeg --libdir=/usr/lib64 --mandir=/usr/share/man --arch=x86_64 --optflags='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic' --enable-bzlib --disable-crystalhd --enable-libass --enable-libdc1394 --enable-libfaac --enable-nonfree --disable-indev=jack --enable-libfreetype --enable-libgsm --enable-libmp3lame --enable-openal --enable-libopencv --enable-libopenjpeg --enable-libopus --enable-librtmp --enable-libtheora --enable-libvorbis --enable-libvpx --enable-libx264 --enable-libxvid --enable-x11grab --enable-avfilter --enable-avresample --enable-postproc --enable-pthreads --disable-static --enable-shared --enable-gpl --disable-debug --disable-stripping --enable-libcaca --shlibdir=/usr/lib64 --enable-runtime-cpudetect
     libavutil      54. 15.100 / 54. 15.100
     libavcodec     56. 13.100 / 56. 13.100
     libavformat    56. 15.102 / 56. 15.102
     libavdevice    56.  3.100 / 56.  3.100
     libavfilter     5.  2.103 /  5.  2.103
     libavresample   2.  1.  0 /  2.  1.  0
     libswscale      3.  1.101 /  3.  1.101
     libswresample   1.  1.100 /  1.  1.100
     libpostproc    53.  3.100 / 53.  3.100
    Metadata:
     Server                NGINX RTMP (github.com/arut/nginx-rtmp-module)
     width                 320.00
     height                240.00
     displayWidth          320.00
     displayHeight         240.00
     duration              0.00
     framerate             0.00
     fps                   0.00
     videodatarate         261.00
     videocodecid          7.00
     audiodatarate         62.00
     audiocodecid          10.00
     profile
     level
    [live_flv @ 0x1edb0820] Could not find codec parameters for stream 0 (Video: none, none, 267 kb/s): unknown codec
    Consider increasing the value for the 'analyzeduration' and 'probesize' options
    Input #0, live_flv, from 'rtmp://<server>/input/<stream>':
     Metadata:
       Server          : NGINX RTMP (github.com/arut/nginx-rtmp-module)
       displayWidth    : 320
       displayHeight   : 240
       fps             : 0
       profile         :
       level           :
     Duration: 00:00:00.00, start: 16.768000, bitrate: N/A
       Stream #0:0: Video: none, none, 267 kb/s, 1k tbr, 1k tbn, 1k tbc
       Stream #0:1: Audio: aac (LC), 16000 Hz, mono, fltp, 63 kb/s
    Unsupported codec with id 0 for input stream 0
    </stream></server>

    I am not, by any means, an expert in H264 or video encoding. I know that the encoded frames that come out from MediaCodec contain SPS NAL, PPS NAL, and frame NAL units. I’ve also written the MediaCodec output into a file and was able to play it back (I did have to specify the format and framerate as otherwise it would play too fast).

    My assumption is that things should work (see how little I know :)). Knowing that SPS and PPS are written out, decoder should know enough. Yet, ffprobe fails to recognize codec, fps, and other video information. Do I need to pass packet flag information to FFmpegFrameRecorder.java:record() function ? Or should I use direct buffer ? Any suggestion will be appreciated ! I should figure things out with a hint.

    PS : I know that some codecs use Planar and other SemiPlanar color formats. That distinction will come later if I get past this. Also, I didn’t go the Surface to MediaCodec way because I need to support API 17 and it requires more changes than this route, which I think helps me understand the more basic flow. Agan, I appreciate any suggestions. Please let me know if something needs to be clarified.

    Update #1

    So having done more testing, I see that my encoder outputs the following frames :

    000000016742800DDA0507E806D0A1350000000168CE06E2
    0000000165B840A6F1E7F0EA24000AE73BEB5F51CC7000233A84240...
    0000000141E2031364E387FD4F9BB3D67F51CC7000279B9F9CFE811...
    0000000141E40304423FFFFF0B7867F89FAFFFFFFFFFFCBE8EF25E6...
    0000000141E602899A3512EF8AEAD1379F0650CC3F905131504F839...
    ...

    The very first frame contains SPS and PPS. From what I was able to see, these are transmitted only once. The rest are NAL types 1 and 5. So, my assumption is that, for ffprobe to see stream info not only when the stream starts, I should capture SPS and PPS frames and re-transmit them myself periodically, after a certain number of frames, or perhaps before every I-frame. What do you think ?

    Update #2

    Unable to validate that I’m writing frames successfully. After having tried to read back the written packet, I cannot validate written bytes. As strange, on successful write of IPL image and streaming, I also cannot print out bytes of encoded packet after avcodec_encode_video2. Hit the official dead end.