Recherche avancée

Médias (1)

Mot : - Tags -/ticket

Autres articles (41)

  • MediaSPIP Core : La Configuration

    9 novembre 2010, par

    MediaSPIP Core fournit par défaut trois pages différentes de configuration (ces pages utilisent le plugin de configuration CFG pour fonctionner) : une page spécifique à la configuration générale du squelettes ; une page spécifique à la configuration de la page d’accueil du site ; une page spécifique à la configuration des secteurs ;
    Il fournit également une page supplémentaire qui n’apparait que lorsque certains plugins sont activés permettant de contrôler l’affichage et les fonctionnalités spécifiques (...)

  • Encoding and processing into web-friendly formats

    13 avril 2011, par

    MediaSPIP automatically converts uploaded files to internet-compatible formats.
    Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
    Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
    Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
    All uploaded files are stored online in their original format, so you can (...)

  • Configuration spécifique d’Apache

    4 février 2011, par

    Modules spécifiques
    Pour la configuration d’Apache, il est conseillé d’activer certains modules non spécifiques à MediaSPIP, mais permettant d’améliorer les performances : mod_deflate et mod_headers pour compresser automatiquement via Apache les pages. Cf ce tutoriel ; mode_expires pour gérer correctement l’expiration des hits. Cf ce tutoriel ;
    Il est également conseillé d’ajouter la prise en charge par apache du mime-type pour les fichiers WebM comme indiqué dans ce tutoriel.
    Création d’un (...)

Sur d’autres sites (5995)

  • Convert input gray16le to RGB output stream

    22 novembre 2018, par Albert Garcia

    I’m reading a v4l2 format input video with pixel format gray16le from a USB camera connected to my computer. I can display it without problems.

    My question is that I want to read that video with ffmpeg (on /dev/video0) and stream it to a websocket. If I read a "normal" rgb video, this is the command that I’m using to accomplish it :

    ffmpeg -f v4l2 -i /dev/video0 -framerate 9 -c:v mpeg1video -f mpegts https://......

    When I try to read from the gray16le format camera, I’ve tried lots of commands but the output is always in grayscale. For example :

    ffmpeg -f v4l2 -i /dev/video0 -framerate 9 -pix_fmt yuyv422 -c:v mpeg1video -f mpegts https://......
    ffmpeg -f v4l2 -i /dev/video0 -framerate 9 -pix_fmt rgb24 -c:v mpeg1video -f mpegts https://......

    Is there an option in ffmpeg to apply a colormap or a LUT to the input video before streaming it ?

    Thanks !

  • Unable to build ; undefined symbols for architecture x86_64 :

    30 août 2021, par switt

    In the process of trying to create an RTSP screen streaming server, I tried looking at CocoaSplit for an example.

    



    Problem is, I can't get it to compile.

    



    I have FFMpeg installed on my computer using Homebrew following FFMpeg's guide on https://trac.ffmpeg.org/wiki/CompilationGuide/MacOSX .

    



    But on compilation, Xcode is throwing out errors such as

    



    Undefined symbols for architecture x86_64:

"_cmnMemAlloc", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
"_cmnMemCheck", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
"_cmnMemCopy", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
"_cmnMemFree", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
"_cmnMemSet", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)


    



    and so on.

    



    The header search paths field in project settings has the directory of /usr/local/include in it.

    



    What seems to be the problem here ?

    


  • Unable to build ; undefined symbols for architecture x86_64 :

    24 mars 2016, par switt

    In the process of trying to create an RTSP screen streaming server, I tried looking at CocoaSplit for an example.

    Problem is, I can’t get it to compile.

    I have FFMpeg installed on my computer using Homebrew following FFMpeg’s guide on https://trac.ffmpeg.org/wiki/CompilationGuide/MacOSX .

    But on compilation, Xcode is throwing out errors such as

    Undefined symbols for architecture x86_64:

    "_cmnMemAlloc", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
    "_cmnMemCheck", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
    "_cmnMemCopy", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
    "_cmnMemFree", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)
    "_cmnMemSet", referenced from:_aac_encode_init in libavcodec.a(libvo-aacenc.o)

    and so on.

    The header search paths field in project settings has the directory of /usr/local/include in it.

    What seems to be the problem here ?