Skip to content
Advertisement

Tag: ffmpeg

Extract the time of the video with ffmpeg in Ubuntu

I am using Ubuntu 14.04.5 LTS 32 bit and ffmpeg to extract the time of a video. In Windows 10 the command works: In Linux ffmpeg returns the error in red: At least one output file must be specified The error complete is here: How can I fix this problem? Answer Use Or you could just use ffprobe

Integrate ffmpeg to unity via plugin OS X and Linux

I need to use a ffmpeg wrapper (so we can use it in unity) for a lecture. Unfortunately the teachers only provide a solution to do this for Windows. They are referring to https://github.com/Ruslan-B/FFmpeg.AutoGen which we should use to generate a wrapper for our OS. That’s where my problems begin: I have downloaded ffmpeg via macport, but I have no

Restart ffmpeg process using monit if TOTAL CPU is less than 1%

I have used kind of similar solution like this Restarting ffmpeg process using monit to restart my ffmpeg stream in case it fails for some reason. Remember its not duplicate problem/question, because I have other issues unlike the example question/solution Restarting ffmpeg process using monit, which I’m gonna explain below. So here is my monit configuration: Here is my streambash.sh

HTTP Live Stream stops playing after a while

I have a problem with streaming with ffserver. After I start ffserver and desktop-capture, everything seems to work fine. Then I open the browser and access the output(http://localhost:8090/test1.mpeg). It plays fine for 6-7 seconds then it stops and I have to refresh the page to get it work again. Does anyone know why that happens and how I can correct

FFmpeg not copying all audio streams

I’m having trouble getting ffmpeg to copy all audio streams from a .mp4 file. After hours of searching online, it appears this should copy all streams (as shown in example 4 here): in.mp4 contains 3 streams: Video Audio track 1 Audio track 2 out.mp4 (which should be identical to in.mp4) contains only 2 streams: Video Audio track 1 FFmpeg does

How can I run command line FFMPEG and accept multiple pipes (video and audio) without blocking on the first input?

I’m trying to mux h264 and aac created with MediaCodec using FFMPEG, and also use FFMPEG’s RTMP support to send to youtube. I’ve created two pipes, and am writing from java (android) through WriteableByteChannels. I can send to one pipe just fine (accepting null audio) like this: YouTube streaming works perfectly (but I have no audio). Using two pipes this

Advertisement