Recherche avancée

Médias (91)

Autres articles (18)

  • Keeping control of your media in your hands

    13 avril 2011, par

    The vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
    While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
    MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
    MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...)

  • Submit bugs and patches

    13 avril 2011

    Unfortunately a software is never perfect.
    If you think you have found a bug, report it using our ticket system. Please to help us to fix it by providing the following information : the browser you are using, including the exact version as precise an explanation as possible of the problem if possible, the steps taken resulting in the problem a link to the site / page in question
    If you think you have solved the bug, fill in a ticket and attach to it a corrective patch.
    You may also (...)

  • Contribute to a better visual interface

    13 avril 2011

    MediaSPIP is based on a system of themes and templates. Templates define the placement of information on the page, and can be adapted to a wide range of uses. Themes define the overall graphic appearance of the site.
    Anyone can submit a new graphic theme or template and make it available to the MediaSPIP community.

Sur d’autres sites (2716)

  • How to set background to subtitle in ffmpeg ?

    10 avril 2017, par supermario

    It is described here how ot burn a srt file into a video.
    However, I want to put a semi-transparent background to the subtitles so that the texts can be read more easily. How can I do that ?

  • timelapse images into a movie, 500 at a time

    2 mars 2017, par molly78

    I am trying to make a script to turn a bunch of timelapse images into a movie, using ffmpeg.

    The latest problem is how to loop thru the images in, say, batches of 500.

    There could be 100 images from the day, or there could be 5000 images.

    The reason for breaking this apart is due to running out of memory.

    Afterwards I would need to cat them using MP4Box to join all together...

    I am entirely new to bash, but not entirely programming.

    What I think needs to happen is this

    1) read in the folders contents as the images may not be consecutively named

    2) send ffmpeg a list of 500 at a time to process (https://trac.ffmpeg.org/wiki/Concatenate)

    2b) while you’re looping thru this, set a counter to determine how many loops you’ve done

    3) use the number of loops to create the MP4Box cat command line to join them all at the end.

    the basic script that works if there’s only say 500 images is :

    #!/bin/bash

    dy=$(date '+%Y-%m-%d')

    ffmpeg -framerate 24 -s hd1080 -pattern_type glob -i "/mnt/cams/Camera1/$dy/*.jpg" -vcodec libx264 -pix_fmt yuv420p Cam1-"$dy".mp4

    MP4Box’s cat command looks like :

    MP4Box -cat Cam1-$dy7.mp4 -cat Cam1-$dy6.mp4 -cat Cam1-$dy5.mp4 -cat Cam1-$dy4.mp4 -cat Cam1-$dy3.mp4 -cat Cam1-$dy2.mp4 -cat Cam1-$dy1.mp4 "Cam1 - $dy1 to $dy7.mp4"

    Needless to say help is immensely appreciated for my project

  • ffmpeg and gnu parallel

    16 août 2013, par souvik

    My work would require me to encode a few thousand movies in a few days. Each movie needs to be encoded in 3 different formats. I use ffmpeg to output these formats in parallel with a single read of the input source as detailed here : http://ffmpeg.org/trac/ffmpeg/wiki/Creating%20multiple%20outputs

    In addition, I am using GNU Parallel to encode from multiple video files in parallel. We have four blade servers of different configurations (48, 32, 16 and 16 cores) encoding videos in parallel. Ideally, we should be able to encode 112 videos in parallel.

    However, it seems that encoding completes faster on machines with lesser cores. I have 16 completed encodes on the 16 core servers in around 4 hours, while it takes close to 10 hours for 48 encodes to complete on the 48 core system. What could be the bottleneck ? A typical encode command is as follows :

    ffmpeg -i sample.mpg -y -vcodec libx264 -vprofile baseline -level 30 -acodec libfdk_aac -ab 128k -ac 2 -b:v 500K -threads 1  encoded/sample_enc.mp4

    Any pointers highly appreciated. Thanks !