Recherche avancée

Médias (91)

Autres articles (60)

  • Websites made ​​with MediaSPIP

    2 mai 2011, par

    This page lists some websites based on MediaSPIP.

  • Les sons

    15 mai 2013, par
  • Creating farms of unique websites

    13 avril 2011, par

    MediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
    This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...)

Sur d’autres sites (5129)

  • Add a side data type for audio service type.

    14 novembre 2014, par Anton Khirnov
    Add a side data type for audio service type.
    

    Currently, audio service type is a field in AVCodecContext. However,
    side data is more appropriate for this kind of information.

    • [DBH] doc/APIchanges
    • [DBH] libavcodec/avcodec.h
    • [DBH] libavcodec/utils.c
    • [DBH] libavcodec/version.h
    • [DBH] libavfilter/af_ashowinfo.c
    • [DBH] libavutil/frame.h
    • [DBH] libavutil/version.h
  • What google cloud service can be used to process files stored in Firebase Cloud Storage with FFmpeg ? [closed]

    1er mai 2021, par uponly

    I am building a ReactJs app and I am trying to figure out a way to process files (images, videos, and audio of any type) that are stored in my Firebase storage bucket using FFmpeg. Currently, I have set up the functionality for allowing the user to upload files to my storage bucket, and a corresponding URL link is stored in a document in Firestore.

    


    Ideally, I'd love to do this with Cloud Functions HTTP triggers because I have all of that setup already. It would be nice to just call an HTTP trigger to process the file after it has been uploaded. However, after a bit of research, my current understanding is to somehow deploy my app using a flexible Google App Engine environment because apparently, it is the only way to set a manual timeout in case I have to process a long, high-quality video, for example. Thus I wouldn't be able to use Cloud Functions because there is a very short timeout period which may lead to the files not being fully processed.

    


    Here is the user flow I am trying to achieve, which may help make things more clear :

    


      

    1. [Done] The user uploads a file (image, audio, or video) to Firebase cloud storage. A URL is also stored in their corresponding user document in Firestore.
    2. 


    3. [Here and the steps onward are what I am trying to achieve] After the file has been stored, I'd like to kick off some sort of function that grabs the newly stored file and begin to process it in the cloud.
    4. 


    5. Store the newly processed file back into the Cloud Storage bucket
    6. 


    7. Allow the user to preview the processed file (by streaming it ideally, if possible).
    8. 


    


    In steps 2 and onward, I am just generally confused about what Google service I should be using to process my file in the cloud with FFmpeg. As well as how I can connect it to my React app, client-side. If I have to go the Google App Engine route, how do I go about connecting app engine to my React App such that I don't have to build my app and deploy it, as my app is still in development ?

    


    This is not a coding question so I apologize if this is the wrong place to post in. I am new to all this, any and all help is greatly appreciated. Thank you.

    


  • Using FFmpegMediaPlayer - Apk size increased by 35MB approx - (without library apk size - 4mb)

    27 janvier 2018, par Arren

    Is there a workaround to this, as the size of the apk is drastically increased and the library was added through normal procedure -
    (compile ’com.github.wseemann:FFmpegMediaPlayer:1.0.4’)

    and not through manual inclusion or downloading I wonder whether anyone else has this problem.