Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I apologize if there is a better place to post this. I recently came across a couple CVS single-use video cameras that I had been using back in 2005. I still have the USB cable I made and now that I've switched to Mac I can use Puretool to extract the old videos from the cameras. The problem I'm having is that I can only hear the audio when viewing the videos with VLC media player. I read on PureTools site, a common problem is that the audio codec used is incompatible with Quicktime and the audio stream needs to be converted. There is a link on their site to a program called FixSound which is purported to fix this, but the link is dead and I've been unable to find any further information about this. If anyone could point me in the right direction, I'd be very grateful. All the videos I'm dealing with are from the first year of my son's life and I'd really like to be able to archive them in a more uniform format. I would think this would be possible to do using Adobe Audition or some other program in Adobe's CS6 Master Collection but I've only dealt with Photoshop, Illustrator, Dreamweaver, Flash & Fireworks. I have no experience with audio or video yet.
Thanks,
Mark
I found a workaround but if anyone has a better suggestion, please post.
I used SoundConverter / GStreamer on my Ubuntu box to convert the audio from the old video to mp3 then combined that with the original video on my iMac using Encore and built it as .f4v (for achiving, I can just build a DVD from Encore). I guess from there I can use Media Encoder to come up with a Quicktime, 3gp, etc... version.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 2 years ago.
Improve this question
For example, how would a washing machine play a specific sound (MP3 or WAV or whatever) when finished? Which external components are needed?
I know this is a really vague question but I could not find any real solutions except some kind of MP3 player module which seems to be more "end-user friendly" instead of being used programmatically (skip and pause buttons).
Various beeps can be made simply by toggling a pin, but I assume your are looking for something more melodious, and without using a sound synthesiser.
For short sound bites: you can store them in (external) memory and play them through a DAC, if your MCU has one. I put a basic implementation of this for SAMD21G-based Arduinos here; the idea is the same for all MCUs that have a DAC. If your MCU doesn't have a DAC, you can use an external one, or make a "poor man's DAC" by using PWM and low-pass filtering the output.
Another option is creating an I2S stream from stored samples and send them to an external amplifier (module) that accepts I2S. It helps when your MCU has an I2S peripheral.
For longer sounds, your best bet is probably playing them from a MCU-controlled MP3-player, because you need the SD-card storage or a similarly large storage anyway for large sound files.
In all cases, you'll need an external amplifier and speaker.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I want to know about Linux audio, i spent a lot of time on reading but i didn't understand(clearly). Can anybody give a brief information on various Linux audio sub systems(Like OSS, ALSA, JACK, Gstreamer, Phonon, Xine)?.
Any help, Thanks in advance.
I once wrote a famous blog post about the jungle of Linux audio output formats. You can find it here.
Regrettably, the picture is no longer there, here's a copy:
It's a bit old (dating from 2007), but I hope it gives you the general idea. OSS and ALSA are the layers closest to the actual audio hardware. All the other libraries and frameworks simply talk to those lower layers. And as you can see, some of these libs and frameworks actually have wrappers around other libs and frameworks.
Which layer you want to call upon depends largely on what you wish to accomplish.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I have a small application I have been working on for some time to help my son with his speech delay. I am using gstreamer to play phonemes, small audio clips about 100ms long.
Things are fine but I find it a bit distasteful to use a library that has such powerful video features for an audio only application.
I was thinking that an audio programing language might be able to play short audio clips on-the-fly too.
I've been reading about these DSLs and there are quite a few of them. However it seems that csound, Supercollider and Chuck are the front runners.
All I really need to do right now is to play small audio clips, preferably from a C binding, in near real time and I only need to run on Linux.
Could anyone help me pick a language for this? There are so many features I "can't see the forest for the trees".
Once I have one picked out, I will have a tool for my immediate needs and a platform to grow with as my needs change.
SuperCollider has great real-time capabilities and makes it pretty easy to play sound files via its sclang interpreter. It's also not that hard to communicate with the supercollider server (scsynth, the part of supercollider that actually synthesize sounds) via open sound control (OSC) messages, so you can control the synth from another, separate application.
Don't know much about chuck but I hear it's good for on-the-fly, live audio programming too so it might work.
I wouldn't recommand CSound since it's meant to be used for composition, more like a "compiled" language vs interpreted. You basically write a score file and generate a whole sound file from that, so it's probably not what you want.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 3 years ago.
Improve this question
I have a requirement where I need to enable playback (full screen) of a h.264 MPEG4 (thanks for the correction!) video from a local network, launchable from a browser link on a Windows workstation, and be frame accurate.
By frame accurate I mean that I need to be able to scrub through the video in the same way you would with a vtr, stop at a frame, and then move backwards and forwards frame by frame (it is for a very specific compliance requirement where have to be able to check every frame if there is something that is potentially against broadcasting guidelines).
The application itself is used to capture notes while viewing the material, so the end model is for a dual monitor workstation, with a web form in one, the video playing full screen in the second (no issue launching the video and manually having to move it to the second screen), and then the user controls the video via keyboard shortcuts or a jog shuttle.
I have looked at QT, but the java bindings seem to be dead or nearly so, flash isn't frame accurate, VLC given its streaming heritage seems to be only able to move forward by a frame and not backwards, and all I have left are commercial offerings that in my experience are difficult and expensive to change.
Any ideas of where I should look or alternative options? Any advice appreciated!
An alternative would be on the server side, dumping each frame out as an image, and then just stepping through the video via images rather than trying to get a video player to work frame by frame.
Here's an ffmpeg command line to dump frames out of a video, to get you started:
ffmpeg -i video.mpeg -f image2 frames/frame-%03d.jpeg
If you wanted to get fancy, you could try using a video player for playing the video, and then when you stop and want to step by frames, replace it with a frame-by-frame view implemented using images. This would take a bit more work to convert video times to frame numbers when you pause the video, but may give you what you need.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 2 years ago.
Improve this question
I'm looking for (linux) command line tools that can help with video editing. I am mostly interested in:
Cutting
Transitions
Effects
Any pointers would be appreciated (I know ffmpeg can do basic cutting, but not much beyond that afaik).
Two tools I use are transcode and mencoder.
Transcode is a suite of command line
utilities for transcoding video and
audio codecs, and for converting
beween different container formats.
A variety of video and audio pre and
post-processing filters are available,
including (but not limited to):
framerate conversion
smoothing
cutting
MEncoder is a free command line video decoding, encoding and filtering tool based on mplayer.
Avisynth under WINE is your best bet. You can use ffmpeg or mencoder to do the actual encoding, and avisynth .avs files for the actual editing. You may have to use mencoder under WINE to access the avs files as well. I haven't tried this myself, but it should work, barring any WINE issues.
If this is part of some production app, if there's any way you can farm the work out to an actual windows computer you'll save yourself a lot of pain.
If you want to get your hands dirty you could try using the gstreamer and Gnonlin to make your own CLI video app. You can actually use a language like Ruby to call gstreamer from, so you don't have to use C.