Recherche avancée

Médias (0)

Mot : - Tags -/diogene

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (16)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Organiser par catégorie

    17 mai 2013, par

    Dans MédiaSPIP, une rubrique a 2 noms : catégorie et rubrique.
    Les différents documents stockés dans MédiaSPIP peuvent être rangés dans différentes catégories. On peut créer une catégorie en cliquant sur "publier une catégorie" dans le menu publier en haut à droite ( après authentification ). Une catégorie peut être rangée dans une autre catégorie aussi ce qui fait qu’on peut construire une arborescence de catégories.
    Lors de la publication prochaine d’un document, la nouvelle catégorie créée sera proposée (...)

  • Les thèmes de MediaSpip

    4 juin 2013

    3 thèmes sont proposés à l’origine par MédiaSPIP. L’utilisateur MédiaSPIP peut rajouter des thèmes selon ses besoins.
    Thèmes MediaSPIP
    3 thèmes ont été développés au départ pour MediaSPIP : * SPIPeo : thème par défaut de MédiaSPIP. Il met en avant la présentation du site et les documents média les plus récents ( le type de tri peut être modifié - titre, popularité, date) . * Arscenic : il s’agit du thème utilisé sur le site officiel du projet, constitué notamment d’un bandeau rouge en début de page. La structure (...)

Sur d’autres sites (4998)

  • How can I concatenate MP4 videos and exactly match this combined video to an audio track in ffmpeg ?

    17 juin 2020, par mattador

    Basically I want to combine 4 videos into a single video, then swap out the concatenated audio for another track. The audio track needs to be unaltered and sync exactly to the video.

    



    Specifically, I'm creating multipaged dynamic audio spectrograms (visualizations of a sound file) using ggplot2 and gganimate in R, which I want to combine using ffmpeg. So the original WAV is chopped up into 3 sec segments, a video is created to visualize frequency changes over time for each segment, which I want to combine into 1 video with the original audio track matched up.

    



    All of my relevant files are at github here in the vids folder : DL link here

    



    The individual dynamic spectrograms export fine (1-4.mp4 in the archive). They are each 20ms too long ; they should be exactly 3s, but otherwise great.

    



    I then use ffmpeg to concatenate them :

    



    ffmpeg -f concat -safe 0 -i "mp4Segments.txt" -codec copy -t 12 -y "(1-4)vid.mp4"


    



    but you can see the end of (1-4)vid.mp4 is getting a little off (the cursor is not all the way to the end, so I guess the video is getting cut off before the end. Not a deal-breaker, since the audio still aligns very well to the video cursor.

    



    Now, since the concatenated video introduces glitchy audio between clips, I want to take the uncut original audio and add that to the assembled video.

    



    ffmpeg -ss 0 -i "(1-4)vid.mp4" -i "origWAV.wav"  -c:v libx264 -map 0:v:0 -map 1:a:0 -c:a aac -ac 1 -b:a 192k -y -vsync 1 -t 12 "(1-4)vid+origWAV.mp4"


    



    but the audio is slightly off now, and ends even further before the end of the video. Since it's really important to align audio and video precisely for this, how can I speed up the video to fit the exact duration of the audio or otherwise fix this issue ?

    


  • movenc : Add a flag for indicating a discontinuous fragment

    20 novembre 2014, par Martin Storsjö
    movenc : Add a flag for indicating a discontinuous fragment
    

    This allows creating a later mp4 fragment without sequentially
    writing the earlier ones before (when called from a segmenter).

    Normally when writing a fragmented mp4 file sequentially, the
    first timestamps of a fragment are adjusted to match the
    end of the previous fragment, to make sure the timestamp is the
    same, even if it is calculated as the sum of previous fragment
    durations. (And for the first packet in a file, the offset of
    the first packet is written using an edit list.)

    When writing an individual mp4 fragment discontinuously like this
    (with potentially writing the earlier fragments separately later),
    there’s a risk of getting a gap in the timeline if the duration
    field of the last packet in the previous fragment doesn’t match up
    with the start time of the next fragment.

    Using this requires setting -avoid_negative_ts make_non_negative
    (or -avoid_negative_ts 0).

    Signed-off-by : Martin Storsjö <martin@martin.st>

    • [DH] libavformat/movenc.c
    • [DH] libavformat/movenc.h
  • CCTV DVR recovery

    5 juin 2014, par bitmadmax

    I need to recover footage from one of these unbranded CCTV DVR units. I am unable to gain access to the unit via a password, factory reset or backdoor, which would otherwise have let me "export" the footage.

    So I have pulled the hard disk and made a copy of it. The original unit is Linux-based which ran off a chip, and an EXT filesystem laid down on the hard disk itself. There is one partition which appears to contain the "raw" CCTV footage.

    The hex structure appears as though each frame begins with 00dcH264, 01dcH264, 10dcH264 and so on... I think the 00, 01, 10 etc refers to the channel/camera input number. The hex contents after the H264 portion appears completely random, of variable length and has no signature, and is followed immediately with the next xxdcH264 header.

    When I use ffprobe across the entire raw footage partition, it is only detecting the 1 video stream. I can use ffmpeg to "force" it to read it in as H.264 and dump it out into an MP4 file that I can play. However, this gives me the problem that the footage as an individual file is quite "noisy" and has frames from each channel appearing randomly.

    Is there anyway I could maybe separate out the footage into say a file for each channel ? I am not a great programmer so something as simple as possible would be ideal. I did note that ffprobe determined that some of the frames were of different resolution - it is likely that the different channels recorded their image at different dimensions, I just wondered if another way might be to get ffprobe/ffmpeg to output frames based on their image dimensions perhaps ?

    EDIT : sample hex below :

    Offset      0  1  2  3  4  5  6  7   8  9 10 11 12 13 14 15
    00000000   32 31 64 63 48 32 36 34  76 08 00 00 00 00 00 00   21dcH264v      
    00000016   1E 8B 77 07 6A 00 00 00  00 00 00 01 61 FC 03 51    ‹w j       aü Q
    00000032   42 83 FC C5 C6 BD 50 63  5F 37 68 B6 B5 B8 4F E0   BƒüÅÆ½Pc_7h¶µ¸Oà
    00000048   2E 74 B8 FF BB BD 70 1C  DE 1E 7A 8C 0F            .t¸ÿ»½p Þ zŒ

    (variable length)....

    Offset      0  1  2  3  4  5  6  7   8  9 10 11 12 13 14 15
    00000000   D3 30 59 01 FA 98 08 A0  0A 27 01 49 19 C1 0C 89   Ó0Y ú˜   ' I Á ‰
    00000016   83 AC 6B 28 03 8D 92 29  38 18 BA 95 82 B3 BA C6   ƒ¬k(  ’)8 º•‚³ºÆ
    00000032   33 0A 74 10 1A D5 89 3B  90 A0 19 1F A9 93 00 10   3 t  Õ‰;    ©“  
    00000048   92 AA 70 32 31 64 63 48  32 36 34                  ’ªp21dcH264