How can I access Kinect audio data with OpenNI? - audio

I am trying to get audio using Kinect's microphones, but can't even even get the sample running.
The sample that comes with OpenNI (Sample-NiAudioSample) exits with this message:
No audio node was found!
Has anyone managed to get read audio from a Kinect?
If so, how?

You can't. Last I checked (a few weeks ago) audio was accessible through the libfreenect drivers (OpenKinect) not OpenNI.

Related

Getting data from audio mixer

I am trying to build an open-source in-ear monitoring system. I have created the UI and was wondering how I would get the channels that are on an audio mixing console so that I can edit the channels and stream them to each musician. Is there a certain protocol that all the mixers use? You can find the project at https://gitlab.com/openstagemix. We would love to have contributors.
I can't really test whether this is the correct answer as I am trapped in my house during the coronavirus time. But, all mixers use something called OSC which is a protocol between mixers, synthesizers, etc. to computers. You can find more information here http://opensoundcontrol.org/introduction-osc.
Update:
It's neither! I am going to use the AES67 standard to receive information from my mixer and with that process the audio. This is because my mixer is ethernet capable.

PulseAudio static noise for 10 minutes after boot

I am building a Bluetooth audio receiver as an embedded system with the CHIP sbc (single board computer) from getchip.com. Pretty similar to Raspberry Pi, runs Debian Jessie, too.
I am using the onboard 3.5mm jack as an audio output. I configured PulseAudio to receive the Bluetooth audio and redirect it to the ALSA sound driver.
Everything works flawlessly except for static noise on the output.
Directly after boot there is a medium loud sum in the few hundred Hz region.
It´s always in the background, even if I play something via bluetooth or locally via CLI.
The interesting part is that it disappears after exactly 10min and 10sec after powerup, so I think exactly 10min after the startup of PulseAudio or ALSA.
I couldn´t find a reason for it.
I tried the tsched=0 fix in /etc/pulse/system.pa
I unloaded the module suspend-on-idle in /etc/pulse/system.pa
And by the way, I´m running PulseAudio in system-mode, as I´m using it as an embedded system and not a multi-user configuration. I hope I get help from you anyway ;-)
Maybe you have an idea where this noise could come from?
It has to be some sort of software configuration issue, otherwise it wouldn´t disappear after exactly 10mins.
I´ll add the PulseAudio and ALSA configuration files later this day.
Thanks in advance!
Fortunately I solved the problem on my own:
The C.H.I.P. from NextThing has a 3.5mm TRRS jack, which does not only output stereo audio but component video as well.
Now if you plug in a standard 3.5mm jack, the ground pin does does interfere with the component video connector.
Thats why there was this humming noise on the audio output. And thats why it disappeared exactly after 10mins, because the screen idle time is 10min, I think.
So I have to admit, that it was indeed not a programming question as it was a connection issue. Thanks anyway for the quick answer!

How implement real time video encoding using Libde265 and Linux

I been reading a lot about H265 encoder but I'm no really sure how to start a C or Python application to encode a video stream in real time using H.265 encoder from libde265, I all ready install the library and I guess I could use opencv to get the input video stream from a usb camera, do anyone has worked in this type of application ?
If you are not particular about using libde265, which I am not aware of, please give a shot with gstreamer. gstreamer has lots of plugins and examples if you are doing some standard tasks like encoding on your stream. It also integrates well with native development.
I have worked on a project similar to yours, where we did H264 encoding and decoding along with few other things on a live camera feed.
Please find the video of a similar application here: https://www.youtube.com/watch?v=JcpkGDpfVU0
Just my two cents!!

Trouble with audio in Windows 8

Recently, my computer has been acting rather strange. All audio is muted and any time I try to watch a Youtube video the player ends up crashing. I can't seem to find anyone who has had a similar problem but I think I may have found some kind of clue: While trying to run a project I made in C++/python for the Panda3d engine I kept on getting these errors:
AL lib: DoReset: Failed to initialize audio client:0x8889000f
:audio(error): alcCreateContext(_device,NULL): Invalid Device
:audio(error): OpenALAudioManager: No open device or context
I have no idea what this means as I am not very knowledgeable of computers.
it is hard to say a lot without more details, but it appears as though the computer is not detecting your audio device. Try right clicking on the computer icon in metro, and selecting manage. Go to the device manager and see if there are any devices that have a yellow exclamation point. if they do, install drivers from the vendor. If not try updating your audio drivers, and restarting.

RTSP live streaming from IP Camera

Is it possible to see the live stream of an IP camera using RTSP ?
Example URL: rtsp://public ip:554/1363e66e.mp4
The encoding is mp4 h.264 baseline profile at 320 x 240 resolution.
I followed the Wiki link here.
But I get the error: Prefetch error -2
When I try to play using real player on the nokia e72, I get the error: 'General: System Error'.
Please let me know what I can do about this.
There are no video players on Ovi store that can play the stream either but I am able to play the stream on VLC on the desktop.
You can stream it using ReaPlayer if you don't have VLC player in Ovi store. See the port address range supported by your IP camera. Try the range of 1024 - 2000. RTSP supports VLC, Quicktime and Real player. Using any of these objects you can stream it.
So I think here is the case,
There are a few different mp4 containers. Standard one wont allow you to wrap a real time data into a mp4 container because mp4 needs to have a field/atom in its header called
MDAT and it has information about the file and its size as well.(which is written after the file is completely encoded. )
So unless you wake that you can not stream live stuff in mp4 UNLESS it is fragmented mp4.
Media Foundation will allow you to do this when windows 8 is out( i got the intel from the msdn forum so I dont know how true it is).
I dont know what ffmpeg/Gstreamer is capable of. Again if this is a commercial product you are working on you might run into some licensing issues with ffmpeg.
Look at webrtc.
I am guessing best bet it to use webm or ogg/theora but I am not sure if theora can do what you want, This is something I am also working on.
Please share your findings
Thanks.

Resources