
Recherche avancée
Médias (3)
-
The Slip - Artworks
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
-
Podcasting Legal guide
16 mai 2011, par
Mis à jour : Mai 2011
Langue : English
Type : Texte
-
Creativecommons informational flyer
16 mai 2011, par
Mis à jour : Juillet 2013
Langue : English
Type : Texte
Autres articles (99)
-
MediaSPIP 0.1 Beta version
25 avril 2011, parMediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
The zip file provided here only contains the sources of MediaSPIP in its standalone version.
To get a working installation, you must manually install all-software dependencies on the server.
If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...) -
Websites made with MediaSPIP
2 mai 2011, parThis page lists some websites based on MediaSPIP.
-
(Dés)Activation de fonctionnalités (plugins)
18 février 2011, parPour gérer l’ajout et la suppression de fonctionnalités supplémentaires (ou plugins), MediaSPIP utilise à partir de la version 0.2 SVP.
SVP permet l’activation facile de plugins depuis l’espace de configuration de MediaSPIP.
Pour y accéder, il suffit de se rendre dans l’espace de configuration puis de se rendre sur la page "Gestion des plugins".
MediaSPIP est fourni par défaut avec l’ensemble des plugins dits "compatibles", ils ont été testés et intégrés afin de fonctionner parfaitement avec chaque (...)
Sur d’autres sites (8889)
-
I want to print HLS files using ffmpeg in aws lambda (python)
14 avril 2021, par 최우선I implemented it through the link(https://aws.amazon.com/ko/blogs/media/processing-user-generated-content-using-aws-lambda-and-ffmpeg/) here, and it works well.


s3_source_bucket = event['Records'][0]['s3']['bucket']['name']
s3_source_key = event['Records'][0]['s3']['object']['key']

s3_source_basename = os.path.splitext(os.path.basename(s3_source_key))[0]
s3_destination_filename = s3_source_basename + ".m3u8"

s3_client = boto3.client('s3')
s3_source_signed_url = s3_client.generate_presigned_url('get_object',
 Params={'Bucket': s3_source_bucket, 'Key': s3_source_key},
 ExpiresIn=SIGNED_URL_TIMEOUT)


ffmpeg_cmd = "/opt/bin/ffmpeg -i \"" + s3_source_signed_url + "\" -codec: copy -start_number 0 -hls_time 10 -hls_list_size 0 -f hls -"
command1 = shlex.split(ffmpeg_cmd)
p1 = subprocess.run(command1, stdout=subprocess.PIPE, stderr=subprocess.PIPE)

resp = s3_client.put_object(Body=p1.stdout, Bucket=S3_DESTINATION_BUCKET, Key=s3_destination_filename)



However, the actual output through ffmpeg is multiple files. For example test.m3u8, test0.ts, test1.ts .....


But when I print p1.stdout, it looks like multiple files (test.m3u8,test0.ts....) are merged into one file.


Is there a way to get the actual output multiple files (test.m3u8,test0.ts......) from p1.stdout ? Please help.


-
Using ffmpeg to assemble images from S3 into a video
10 juillet 2020, par Mass Dot NetI can easily assemble images from local disk into a video using ffmpeg and passing a
%06d
filespec. Here's what a typical (pseudocode) command would look like :

ffmpeg.exe -hide_banner -y -r 60 -t 12 -i /JpgsToCombine/%06d.JPG <..etc..>



However, I'm struggling to do the same with images stored in AWS S3, without using some third party software to mount a virtual drive (e.g. TNTDrive). The S3 folder containing our images is too large to download to the 20GB ephemeral storage provided for AWS containers, and we're trying to avoid EFS because we'd have to provision expensive bandwidth.


Here's what the HTTP and S3 URLs to each of our JPGs looks like :


# HTTP URL
https://massdotnet.s3.amazonaws.com/jpgs-to-combine/000000.JPG # frame 0
https://massdotnet.s3.amazonaws.com/jpgs-to-combine/000012.JPG # frame 12
https://massdotnet.s3.amazonaws.com/jpgs-to-combine/000123.JPG # frame 123
https://massdotnet.s3.amazonaws.com/jpgs-to-combine/456789.JPG # frame 456789

# S3 URL
s3://massdotnet/jpgs-to-combine/000000.JPG # frame 0
s3://massdotnet/jpgs-to-combine/000012.JPG # frame 12
s3://massdotnet/jpgs-to-combine/000123.JPG # frame 123
s3://massdotnet/jpgs-to-combine/456789.JPG # frame 456789



Is there any way to get ffmpeg to assemble these ? We could generate a signed URL for each S3 file, and put several thousand of those URLs onto a command line with an FFMPEG concat filter. However, we'd run up into the command line input limit in Linux at some point using this approach. I'm hoping there's a better way...


-
AWS Lambda : "Unzipped size must be smaller than 106534017 bytes" after adding single file
17 septembre 2023, par leonWhen trying to deploy my lambdas using AWS through the serverless framework I had no problems until I tried adding the ffmpeg binary.


Now the ffmpeg binaries I have tried to add have ranged from 26 mb to 50 mb. Whichever I add, I get the following error :


UPDATE_FAILED: WhatsappDocumentHandlerLambdaFunction (AWS::Lambda::Function)
Resource handler returned message: "Unzipped size must be smaller than 106534017 bytes (Service: Lambda, Status Code: 400, Request ID: ...)" (RequestToken: ..., HandlerErrorCode: InvalidRequest)



The problem is that I did not add the file to this function. I added it to a completely different one.


I have tried the following things :


- 

- Creating an "empty" function that only contains the ffmpeg binary and a function handler
- Creating a layer that only contains the ffmpeg binary
- Deleting the ffmpeg binary (the error goes away and deployment succeeds
- Varying sizes of ffmpeg binaries between 26 and 50mb
- Getting the ffmpeg-lambda-layer (https://github.com/serverlesspub/ffmpeg-aws-lambda-layer ; https://serverlessrepo.aws.amazon.com/applications/us-east-1/145266761615/ffmpeg-lambda-layer) and deploying it myself












When trying every single one of these options I get the UPDATE_FAILED error in a different function that surely is not too big.


I know I can deploy using a docker image but why complicate things with docker images when it should work ?


I am very thankful for any ideas.