
Recherche avancée
Autres articles (93)
-
Le profil des utilisateurs
12 avril 2011, parChaque utilisateur dispose d’une page de profil lui permettant de modifier ses informations personnelle. Dans le menu de haut de page par défaut, un élément de menu est automatiquement créé à l’initialisation de MediaSPIP, visible uniquement si le visiteur est identifié sur le site.
L’utilisateur a accès à la modification de profil depuis sa page auteur, un lien dans la navigation "Modifier votre profil" est (...) -
Les sons
15 mai 2013, par -
Configurer la prise en compte des langues
15 novembre 2010, parAccéder à la configuration et ajouter des langues prises en compte
Afin de configurer la prise en compte de nouvelles langues, il est nécessaire de se rendre dans la partie "Administrer" du site.
De là, dans le menu de navigation, vous pouvez accéder à une partie "Gestion des langues" permettant d’activer la prise en compte de nouvelles langues.
Chaque nouvelle langue ajoutée reste désactivable tant qu’aucun objet n’est créé dans cette langue. Dans ce cas, elle devient grisée dans la configuration et (...)
Sur d’autres sites (6077)
-
Synchronization Error in Gstreamer/H264 encoding
21 juin 2020, par RyanTrying to produce a frame synced recording platform w/multiple sensors all capturing at the same fps.


Currently can get up to 8 usb cameras streaming and capturing h264 files. These files are reporting same durations, same number of frames, same frame rate etc. I am experiencing some drift in the videos however.


I assume I am not able to write frames to disk fast enough and this results in the shift. However what I do not understand what is 'masking' this, and why it would report the same nb_frames and durations even though real world time is clearly different. The shift can be relatively minor (5-10 frames over 2 minutes) or more extreme (20-30). But what I am looking for is a multicam frame level sync. So each video can produce the exact same frame and a given timestamp.


I know there are many elements at play here. What I'm looking for first is a better understanding of what my exact problem may be and how I can understand it.


-
Is it safe to set AVPacket::duration to one ?
9 septembre 2021, par Vivekanand VI am new to the world of programming of multimedia applications. I just managed to get a hold of FFMPEG's libraries to perform video encoding and decoding in C/C++ and I feel that I have a lack of understanding in many aspects of the whole process.


My application converts an RGB24 array referenced by
unsigned char*
to a AVFrame, and encodes it to a video file.

During initialization, I set the time base in codec context to 1/fps, where fps is the frame rate.


AVStream::time_base
is also set to the same value.

Is there is a mathematical relationship between these time bases to
AV_TIME_BASE
constant ?

How should
AVFrame::pts
be set. Here I am just incrementing by one for every encoding operation.

How should AVPacket::duration be set, before calling
av_interleaved_write_frame
?

What is the significance of
AVPacket::pts
andAVPacket::dts
?

-
Using FFmpeg native libraries with Android-NDK
24 octobre 2013, par ssrpI could write a simple "Hello World" app and got it to compile and run ok, I know how to setup basic structure of an Android project to use native libraries. Could any one tell me how to use FFmpeg libraries to decode a video ;
- How to write a .cpp file to decode a video file using the libraries(Open, Decode, Save)
- How to setup source files(in my case .cpp files) with
Android.mk
file
Please could anyone help.