I am using Ubuntu 14.04.5 LTS 32 bit and ffmpeg to extract the time of a video. In Windows 10 the command works: In Linux ffmpeg returns the error in red: At least one output file must be specified The error complete is here: How can I fix this problem? Answer Use Or you could just use ffprobe
Tag: ffmpeg
Integrate ffmpeg to unity via plugin OS X and Linux
I need to use a ffmpeg wrapper (so we can use it in unity) for a lecture. Unfortunately the teachers only provide a solution to do this for Windows. They are referring to https://github.com/Ruslan-B/FFmpeg.AutoGen which we should use to generate a wrapper for our OS. That’s where my problems begin: I have downloaded ffmpeg via macport, but I have no
Restart ffmpeg process using monit if TOTAL CPU is less than 1%
I have used kind of similar solution like this Restarting ffmpeg process using monit to restart my ffmpeg stream in case it fails for some reason. Remember its not duplicate problem/question, because I have other issues unlike the example question/solution Restarting ffmpeg process using monit, which I’m gonna explain below. So here is my monit configuration: Here is my streambash.sh
How can low frame rate video be made to look more smooth?
I am trying to clean up a video that was recorded in 2003 in low-light conditions on what was possibly a cameraphone. The video has been cleaned up somewhat (cropped, logos removed and stabilized), but it remains quite jerky, due in large part to its low frame rate. What are some tricks that might clean up the video in this
Adjust volume of both input mp3 files while merging using ffmpeg
I am using the following command to merge two audio files (mp3) into one output.mp3 Kindly suggest me how to adjust volume level of both input files to some specific level. I have found the following filter variable but don’t exactly know how to adjust into my command. Any help will be much appreciated.Thanks Answer You would use
HTTP Live Stream stops playing after a while
I have a problem with streaming with ffserver. After I start ffserver and desktop-capture, everything seems to work fine. Then I open the browser and access the output(http://localhost:8090/test1.mpeg). It plays fine for 6-7 seconds then it stops and I have to refresh the page to get it work again. Does anyone know why that happens and how I can correct
FFmpeg not copying all audio streams
I’m having trouble getting ffmpeg to copy all audio streams from a .mp4 file. After hours of searching online, it appears this should copy all streams (as shown in example 4 here): in.mp4 contains 3 streams: Video Audio track 1 Audio track 2 out.mp4 (which should be identical to in.mp4) contains only 2 streams: Video Audio track 1 FFmpeg does
Combining videos with ffmpeg using crossfades and plain cuts
I am writing a script to combine/slice arbitrary video files from S3 into one output video. So far, I am doing this by first trimming the videos to their proper length using ffmpeg -i input-X.mp4 -ss start -t duration slice-X.mp4 and recombining the resultant slices with the ffmpeg concat filter. I want to be able to crossfade and cut between
Concat multiple (self-generated) videos using ffmpeg on raspbian linux
I am a very talented sleep talker, so I decided to write a solution that records the things I talk at night to make funny videos with subtitles of it. The project is nearly done, but I got a big problem with concating videos I generated before. The video parts are generated from single png frames using this command: Then
How can I run command line FFMPEG and accept multiple pipes (video and audio) without blocking on the first input?
I’m trying to mux h264 and aac created with MediaCodec using FFMPEG, and also use FFMPEG’s RTMP support to send to youtube. I’ve created two pipes, and am writing from java (android) through WriteableByteChannels. I can send to one pipe just fine (accepting null audio) like this: YouTube streaming works perfectly (but I have no audio). Using two pipes this