
Recherche avancée
Médias (91)
-
Géodiversité
9 septembre 2011, par ,
Mis à jour : Août 2018
Langue : français
Type : Texte
-
USGS Real-time Earthquakes
8 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
SWFUpload Process
6 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
La conservation du net art au musée. Les stratégies à l’œuvre
26 mai 2011
Mis à jour : Juillet 2013
Langue : français
Type : Texte
-
Podcasting Legal guide
16 mai 2011, par
Mis à jour : Mai 2011
Langue : English
Type : Texte
-
Creativecommons informational flyer
16 mai 2011, par
Mis à jour : Juillet 2013
Langue : English
Type : Texte
Autres articles (83)
-
Contribute to a better visual interface
13 avril 2011MediaSPIP is based on a system of themes and templates. Templates define the placement of information on the page, and can be adapted to a wide range of uses. Themes define the overall graphic appearance of the site.
Anyone can submit a new graphic theme or template and make it available to the MediaSPIP community. -
Configuration spécifique pour PHP5
4 février 2011, parPHP5 est obligatoire, vous pouvez l’installer en suivant ce tutoriel spécifique.
Il est recommandé dans un premier temps de désactiver le safe_mode, cependant, s’il est correctement configuré et que les binaires nécessaires sont accessibles, MediaSPIP devrait fonctionner correctement avec le safe_mode activé.
Modules spécifiques
Il est nécessaire d’installer certains modules PHP spécifiques, via le gestionnaire de paquet de votre distribution ou manuellement : php5-mysql pour la connectivité avec la (...) -
ANNEXE : Les plugins utilisés spécifiquement pour la ferme
5 mars 2010, parLe site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)
Sur d’autres sites (5804)
-
Android - How can I pass camera stream to ffmpeg, using Camera2 library ?
29 mars 2021, par Juan José CetraroI am trying to create an app that shows the camera of the device on the screen, and also streams the camera by srt. To do this, I am using Camera2 library, and ffmpeg (in partucular I am using https://github.com/tanersener/mobile-ffmpeg, that is a ffmpeg wrapper for Android).


My plan is to get the camera stream using Camera2 (using the method onImageAvailable on ImageReader.OnImageAvailableListener class), and send this stream to udp ://localhost:1234. Then, I can use ffmpeg to get that stream by udp, and send it by srt.


I've already solved the part of sending the stream by srt using ffmpeg, and it works fine. In fact, if I set "android_camera" as the input of my ffmpeg command, my app works ok. The problem with this approach, is that if I do that, I block the access to the camera, so I can't show the camera on the screen with another library.


I also found a code that uses Camera2 to stream the camera by udp, and it works, but the problem with this code is that converts each frame to bitmap before sending it by udp, and it makes that it is not performant.


So, I need to know which is the best way to pass the data by udp to ffmpeg, so ffmpeg could process it and send it by srt ?


Camera2 let me to configure which format I want to receive the frames on my listener :


ImageReader.newInstance(1280, 720, ImageFormat.JPEG, /*maxImages*/2);



In this example I am setting JPEG as ImageFormat, but here I let you all the available formats I could use :


UNKNOWN, RGB_565, YV12, Y8, Y16, NV16, NV21, YUY2, JPEG, DEPTH_JPEG, YUV_420_888, YUV_422_888, YUV_444_888, FLEX_RGB_888, FLEX_RGBA_8888, RAW_SENSOR, RAW_PRIVATE, RAW10, RAW12, DEPTH16, DEPTH_POINT_CLOUD, RAW_DEPTH, PRIVATE, HEIC


This is the method where I am going to receive each frame, and what I need to know is what kind of transformation I have to do before sending the frame by udp to ffmpeg ? :


@Override
public void onImageAvailable(ImageReader reader) {

}



Thanks in advance for reading the question :)


-
ExoPlayer and deinterlacing feature
21 février 2021, par SuppamanI want to use ExoPlayer to show interlaced video but, especially for SD content, the video looks really bad without any deinterlacing feature. Looking in the ExoPlayer project it seem the player doesn't have such deinterlacing feature and doesn't seem in plan to be added. I'm thinking to try add deinterlacing by using the deinterlacing engine of ffmpeg library. However it don't seem and easy task. I found some projects that could help me like this ExoPlayerFilter that apply a filter on the frame before show. This could be a possible way to "deinterlace" the frame before show. Another possible suggestion comes from the tread in github Exoplayer project here. This could be another possible way to proceed. My problem is that I have a limited time to try add this feature and I would to know from people having more experience than me what would be most "suggested" way to follow or if someone know another better way to reach my results.


Thank you


-
How to take metadata from .mp3 file and put it to a video as a text using FFmpeg ?
6 décembre 2020, par GrrzlyIn my previously opened topic :


How to make FFmpeg automatically inject mp3 audio tracks in the single cycled muted video



I've got detailed explanation from @llogan how to broadcast looped short muted video on youtube automatically injecting audio tracks in it without interrupting a translation.


I plan to enhance the flow and the next question I faced with is how to dynamically put an additional text to the broadcast.


Prerequisites :


- 

- youtube broadcast is up and running by ffmpeg
- short 3 min video is paying in infinity loop
- audio tracks from playlist are automatically taken by "ffmpeg concat" and injected in the video one by one








this is a basic command to start translation :




ffmpeg -re -fflags +genpts -stream_loop -1 -i video.mp4 -re -f concat
-i input.txt -map 0:v -map 1:a -c:v libx264 -tune stillimage -vf format=yuv420p -c:a copy -g 20 -b:v 2000k -maxrate 2000k -bufsize
8000k -f flv rtmp ://a.rtmp.youtube.com/live2/my-key




Improvements I want to bring


- 

- I plan to store some metadata in audio files (basically it's an artist name and a song name)
- At the moment a particular song starts playing artist/song name should be taken from metadata and displayed on the video as text during the whole song is playing.
- When the current song finishes and a new one starts playing the previous artist/song text should be replaced with the new one etc








My question is how to properly take metadata and add it to the existing broadcast config using ffmpeg ?