
Recherche avancée
Médias (1)
-
Bug de détection d’ogg
22 mars 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Video
Autres articles (38)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Contribute to translation
13 avril 2011You can help us to improve the language used in the software interface to make MediaSPIP more accessible and user-friendly. You can also translate the interface into any language that allows it to spread to new linguistic communities.
To do this, we use the translation interface of SPIP where the all the language modules of MediaSPIP are available. Just subscribe to the mailing list and request further informantion on translation.
MediaSPIP is currently available in French and English (...) -
La file d’attente de SPIPmotion
28 novembre 2010, parUne file d’attente stockée dans la base de donnée
Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)
Sur d’autres sites (3859)
-
Is it possible to pipe an ffmpeg output with multiple files (HLS or DASH)
29 août 2020, par New DevI'm using FFmpeg to generate fragmented MP4s in a dash and HLS format.


ffmpeg -i input.mov -f dash -seg_duration 6 -hls_playlist true output.mbd



The above (simplified) command outputs multiple files, in addition to
output.mdb
(e.g. actual segments,master.m3u8
, etc...)

Is there a way to get each produced file into their individual and separate output streams ?



Broader context :


I'm trying to build a transcoder in Node.js running in Google Cloud, with the idea being that it writes directly to a Google Storage through a writable stream. I can only create a stream per file, but since the number of files is dynamic, I'm not sure how to obtain a stream from each file.


-
How to detect silence in a pyAV AudioFrame ?
29 janvier 2024, par Sachin DoleI want to process streaming audio (coming in from a person speaking on the peer of a webRTC peer connection) to detect when the person is done talking. I have got the audio track and access to individual frames. I see that each frame can be converted to an nd_array using Frame.to_ndarray. I can also see values in the ndarray changing depending on what the person is speaking, what pitch, what volume etc. Now, I want to detect silence on the stream. My question is what is in the ndarray and how can I make sense of the data ?


while True:
 try:
 frame:AudioFrame = await track.recv()
 frame_nd_array = frame.to_ndarray() 



Where can I learn what is in the frame_nd_array ?


-
How to batch auto-generate videos according to audio length
16 septembre 2020, par yocibaI have hundreds of folders of Spanish short stories, with each folder having dozens of individual .MP3 files for each dialogue spoken. Along with the .MP3 files, lies in the folder, a JSON file of the format
["folder1-dialogue-1", "folder1-dialogue-2",...]
.

I want to generate images, each containing the text values of the JSON file, like
folder1-dialogue-1
and generate videos combining that image with its corresponding audio(folder1-dialogue-1.mp3)
, and then merge all the videos in the folder into one -folder1.mp4


How do I go about doing this ?


P.S I'm not straight up asking for code, but just advice as to how I go about doing this.