Recherche avancée

Médias (0)

Mot : - Tags -/clipboard

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (34)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Utilisation et configuration du script

    19 janvier 2011, par

    Informations spécifiques à la distribution Debian
    Si vous utilisez cette distribution, vous devrez activer les dépôts "debian-multimedia" comme expliqué ici :
    Depuis la version 0.3.1 du script, le dépôt peut être automatiquement activé à la suite d’une question.
    Récupération du script
    Le script d’installation peut être récupéré de deux manières différentes.
    Via svn en utilisant la commande pour récupérer le code source à jour :
    svn co (...)

  • Installation en mode standalone

    4 février 2011, par

    L’installation de la distribution MediaSPIP se fait en plusieurs étapes : la récupération des fichiers nécessaires. À ce moment là deux méthodes sont possibles : en installant l’archive ZIP contenant l’ensemble de la distribution ; via SVN en récupérant les sources de chaque modules séparément ; la préconfiguration ; l’installation définitive ;
    [mediaspip_zip]Installation de l’archive ZIP de MediaSPIP
    Ce mode d’installation est la méthode la plus simple afin d’installer l’ensemble de la distribution (...)

Sur d’autres sites (2663)

  • GoPro (MP4) video timestamp sync with precision of milliseconds

    3 février 2021, par Raphael Ottoni

    I need your help with a data sync problem... I m currently trying to sync my GoPro video with real world time (a.k.a my notebook). I manage to sync date and time of my notebook and my GoPro 3+ black perfectly. The problem is that when the GoPro save the files in disk it round up the milliseconds on the creation_time (the milliseconds is always 000000) . Thus, turning the perfect sync impossible. In attachment is a picture of the meta information (extracted by ffprobe) of the MP4 video.

    



    My question is : What I have to do, so the GoPro actually save the creation_time with precision of milliseconds ?

    



    Another small question : Looking at the attachment figure, we see the "timecode" which is a time synchronization data in the format of hours:minuts:seconds:frame. I was thinking that I could use the "frame" value to calculate the missing milliseconds value. If we take this attachment, as a example, we can see that the frame value is "36". Meaning that the millisecond that it started to record was the one associated with the 36th frame of the FPS (in this video : 60fps) value : Some thing like 1000/60 * 36 which is 600 milliseconds, thus the actual creation_time of this video would be : 2017-07-19T18:10:34.600

    



    Is this logic right ? it didn't work ! I don't know what else to do.

    



    P.S : I need this kind of time precision because I will sync the video frames with a external sensor data that is recorded at 11hz.

    



    Please Help

    



    enter image description here

    



    update

    



    I forgot to mention, even if you check the original raw file information, inside the GoPro SSD card, using "stats" to read the creation time (see attachment) it still has the same timestamp without milliseconds.

    



    enter image description here

    


  • avcodec : add AV_CODEC_ID_IVTV_VBI

    14 décembre 2024, par Scott Theisen
    avcodec : add AV_CODEC_ID_IVTV_VBI
    

    IVTV, a Linux driver for TV tuners, and V4L2 utilize
    a coding named after IVTV to carry sliced VBI data
    in MPEG streams produced by tuner cards with
    VBI capture capability and an MPEG-2 encoder SoC.

    IVTV or V4L2 driver will transport the coded data into a
    MPEG-PS private stream ("IVTV") that can be captured
    from the card alongside the video/audio.

    The data could include :
    EIA-608, Teletext, WSS (PAL widescreen signaling),
    or VPS (PAL VCR signaling).

    Signed-off-by : Marth64 <marth64@proxyid.net>

    • [DH] doc/APIchanges
    • [DH] libavcodec/codec_desc.c
    • [DH] libavcodec/codec_id.h
    • [DH] libavcodec/version.c
  • Android : Recording and Streaming at the same time

    29 mars 2016, par Bruno Siqueira

    This is not really a question as much as it is a presentation of all my attempts to solve one of the most challenging functionalities I was faced with.

    I use libstreaming library to stream realtime videos to Wowza Server and I need to record it at the same time inside the SD card. I am presenting below all my attempts in order to collect new ideias from the community.

    Copy bytes from libstreaming stream to a mp4 file

    Development

    We created an interception in libstreaming library to copy all the sent bytes to a mp4 file. Libstreaming sends the bytes to Wowza server through a LocalSocket. It users MediaRecorder to access the camera and the mic of the device and sets the output file as the LocalSocket’s input stream. What we do is create a wrapper around this input stream extending from InputStream and create a File output stream inside it. So, every time libstreaming executes a reading over the LocaSocket’s input stream, we copy all the data to the output stream, trying to create a valid MP4 file.

    Impediment

    When we tried to read the file, it is corrupted. We realized that there are meta information missing from the MP4 file. Specifically the moov atom. We tried to delay the closing of the streaming in order to give time to send this header (this was still a guessing) but it didn’t work. To test the coherence of this data, we used a paid software to try to recover the video, including the header. It became playable, but it was mostly green screen. So this became an not trustable solution. We also tried using "untrunc", a free open source command line program and it couldn’t even start the recovery, since there was no moov atom.

    Use ffmpeg compiled to android to access the camera

    Development

    FFMPEG has a gradle plugin with a java interface to use it inside Android apps. We thought we could access the camera via command line (it is probably in "/dev/video0") and sent it to the media server.

    Impediment

    We got the error "Permission Denied" when trying to access the camera. The workaround would be to root the device to have access to it, but it make the phones loose their warranty and could brick them.

    Use ffmpeg compiled to android combined with MediaRecorder

    Development

    We tried to make FFMPEG stream a mp4 file being recorded inside the phone via MediaRecorder

    Impediment

    FFMPEG can not stream MP4 files that are not yet done with the recording.

    Use ffmpeg compiled to android with libstreaming

    Development

    Libstreaming uses LocalServerSocket as the connection between the app and the server, so we thought that we could use ffmpeg connected with LocalServerSocket local address to copy the streaming directly to a local file inside the SD card. Right after the streaming started, we also ran the ffmpeg command to start recording the data to a file. Using ffmpeg, we believed that it would create a MP4 file in the proper way, which means with the moov atom header included.

    Impediment

    The "address" created is not readable via command line, as a local address inside the phone. So the copy is not possible.

    Use OpenCV

    Development

    OpenCV is an open-source, cross-platform library that provides building blocks for computer vision experiments and applications. It offers high-level interfaces for capturing, processing, and presenting image data. It has their own APIs to connect with the device camera so we started studding it to see if it had the necessary functionalities to stream and record at the same time.

    Impediment

    We found out that the library is not really defined to do this, but more as image mathematical manipulation. We got even the recommendation to use libstreaming (which we do already).

    Use Kickflip SDK

    Development

    Kickflip is a media streaming service that provides their own SDK for development in android and IOS. It also uses HLS instead of RTMP, which is a newer protocol.

    Impediment

    Their SDK requires that we create a Activity with camera view that occupies the entire screen of the device, breaking the usability of our app.

    Use Adobe Air

    Development

    We started consulting other developers of app’s already available in the Play Store, that stream to servers already.

    Impediment

    Getting in touch with those developers, they reassured that would not be possible to record and stream at the same time using this technology. What’s more, we would have to redo the entire app from scratch using Adobe Air.

    UPDATE

    Webrtc

    Development

    We started using WebRTC following this great project. We included the signaling server in our NODEJS server and started doing the standard handshake via socket. We were still toggling between local recording and streaming via webrtc.

    Impediment

    Webrtc does not work in every network configuration. Other than that, the camera acquirement is all native code, which makes a lot harder to try to copy the bytes or intercept it.