Recherche avancée

Médias (91)

Autres articles (101)

  • Use, discuss, criticize

    13 avril 2011, par

    Talk to people directly involved in MediaSPIP’s development, or to people around you who could use MediaSPIP to share, enhance or develop their creative projects.
    The bigger the community, the more MediaSPIP’s potential will be explored and the faster the software will evolve.
    A discussion list is available for all exchanges between users.

  • Le plugin : Podcasts.

    14 juillet 2010, par

    Le problème du podcasting est à nouveau un problème révélateur de la normalisation des transports de données sur Internet.
    Deux formats intéressants existent : Celui développé par Apple, très axé sur l’utilisation d’iTunes dont la SPEC est ici ; Le format "Media RSS Module" qui est plus "libre" notamment soutenu par Yahoo et le logiciel Miro ;
    Types de fichiers supportés dans les flux
    Le format d’Apple n’autorise que les formats suivants dans ses flux : .mp3 audio/mpeg .m4a audio/x-m4a .mp4 (...)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

Sur d’autres sites (5337)

  • avcodec/mjpegdec : Use correct number of codes when init default VLCs

    8 octobre 2020, par Andreas Rheinhardt
    avcodec/mjpegdec : Use correct number of codes when init default VLCs
    

    Commit bbc0d0c1fe2b7ecdc4367295594f084f85ad22f5 made the mjpeg decoder
    use default Huffman tables when none are given, yet when initializing
    the default Huffman tables, it did not use the correct number of entries
    of the arrays used to initialize the tables, but instead it used the
    biggest entry + 1 (as if it were a continuous array 0..biggest entry).
    This worked because the ff_init_vlc_sparse() (and its predecessors)
    always skipped entries with a length of zero and the length of the
    corresponding elements was always initialized to zero with only the
    sizes of the actually existing elements being set to a size > 0 lateron.

    Yet since commit 1249698e1b424cff8e77e6a83cfdbc9d11e01aa7 this is no
    longer so, as build_vlc() actually read the array containing the values
    itself. This implies that the wrong length now leads to a read beyond
    the end of the given array ; this could lead to crashs (but usually
    doesn't) ; it is detectable by ASAN* and this commit fixes it.

    * : AddressSanitizer : global-buffer-overflow on address xy
    ...
    xy is located 0 bytes to the right of global variable 'avpriv_mjpeg_val_ac_luminance'

    Reviewed-by : Paul B Mahol <onemda@gmail.com>
    Signed-off-by : Andreas Rheinhardt <andreas.rheinhardt@gmail.com>

    • [DH] libavcodec/mjpegdec.c
  • Real-time streaming of rawdata images to Android with FFMPEG

    29 juillet 2015, par Philies

    CONDITIONS

    I have an C++ server (Linux) and want to transmit rawdata images (RGB, 32bit) to an Android device in real-time.
    The server generates the rawdata images (with or without bitmap header) continuously every x miliseconds. Now, I want to put the rawdata images into a stream and transmit them without much delay to the Android client.

    I’ve chosen FFMPEG for this kind of job.

    The input for FFMPEG should be the rawdata images, which are generated just in time. The output should be an rtsp stream (h264 or is another format better ?).
    On client side I will play the stream with the Android MediaPlayer. That works for a RTSP url like rtsp ://184.72.239.149/vod/mp4:BigBuckBunny_115k.mov


    PREVIOUS APPROCH :

    I’ve installed FFMPEG on my server.

    Depending on this question : https://ffmpeg.org/pipermail/ffmpeg-user/2013-April/014617.html ... here is my FFMEG command :

    ffmpeg -an -f rawvideo -vcodec rawvideo -pix_fmt rgb32 -r 10 -i -vcodec libx264 -r 30 -tune zerolatency -preset ultrafast -bsf:v h264_mp4toannexb -f mpegts udp://192.168.1.20:1234

    Explanation of the parameters :

    • -an = ignore audio
    • -report = log file in current directory

    INPUT :

    • -f = rawvideo (video format)
    • -vcodec = rawvideo (video codec)
    • -pix_fmt = rgb32
    • -r = 10 (frame rate)

    OUTPUT :

    • -vcodec = libx264
    • -r = 30 (frame rate)
    • -bsf:v = h264_mp4toannexb
    • -f = mpegts udp ://192.168.1.20:1234

    C++ Server Code Snippet :

    /* -------  Get image rawdata from source ------- */
    ...
    /* -------  Create image header which fits to the image rawdata ------- */
    ...

    /* ------- Store the picture local (not necessary?) ------- */
    FILE *f;
    f = fopen("/home/philies/test.bmp","wb");

    //Write Bitmap Headers
    fwrite(&amp;imageFileHeader,1,sizeof(imageFileHeader),f);
    fwrite(imageInfoHeader,1, imageInfoHeader,f);

    //Write Bitmap Rawdata
    fwrite(lastImage.GetBitmapRawData(),1,imageSize,f);


    —> ffmpeg ???

    PROBLEMS/QUESTIONS :

    1. How can I specify the rawvideo of the ffmpeg command respectively how can I specify my bitmap rawdata (with or without header ?) as the input of the FFMPEG ?

    2. Is the FFMPEG command correct ?

    If I fire the FFMPEG command in the terminal I get this error :

    vcodec : no such file or directory

    EDIT :
    In the first step, I will open the udp stream with the VLC player on my Android device. After that I will set up a FFserver to create a RTSP stream

  • Working way to make video from images in C#

    29 août 2011, par Jim Mischel

    Does anybody have a known reliable way to create a video from a series of image files ? Before you mod me down for not searching for the answer before posting the question, and before you fire off a simple message like "use FFMPEG," read the rest of this message.

    I'm trying to create a video, it doesn't matter too much what format as long as it's widely supported, from a series of images (.jpg, .bmp, etc.). My platform is Windows Server 2008, 64-bit. If I can make the video from within my C# program, that's great, but I'm not averse to writing a series of image files to a directory and then firing off an external program to make a video from those images.

    The only constraints are : it must work on my Windows Server 2008 system, and be scriptable. That is, no GUI programs that require operator intervention.

    I found a number of similar questions on StackOverflow, and have tried several of the solutions, all with varying degrees of frustration and none with anything like success.

    FFMPEG looks like a great program. Maybe it is, on Linux. The two Windows builds I downloaded are broken. Given this command line :

     ffmpeg -r 1 -f image2 -i jpeg\*.jpg video.avi

    One of the builds reads the images and then crashes due to data execution prevention. The other reads the first file and then spits out an error message that says "cannot find suitable codec for file jpeg/image2.jpg". Helpful, that. In any case, FFMPEG looks like a non-starter under Windows.

    One answer to a previous posting recommended Splicer . It looks like pretty good code. I compiled the samples and tried to run, but got some cryptic error message about a file not found. It looks like a COM class isn't registered. I suppose I need to install something (DirectShow, maybe, although I thought that was already installed ?). Depending on what's required, I might have a difficult time justifying its installation on a server. ("What ? Why do you need that on a server ?")

    Another answer suggested the AviFile library from Code Project. That looks simple enough : a wrapper around the Windows AviFile subsystem. Except that the AVI files the package creates appear to have all of the frames, but only the first frame shows when I play the AVI in Windows Media Player. Well, that and if you try to create a compressed video, the program throws an exception.

    So, I'm left wondering if there is a good, reliable way to do what I want : on a Windows system, create an AVI or other common video file format from a series of images, either through a .NET API or using an external program. Any help ?