Search Results

Search found 5073 results on 203 pages for 'audio jack'.

Page 3/203 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • Linux: how to use Jellyfish from Jack Meterbridge?

    - by klox
    dear all, i have installed Meterbridge. But,i'm just need to use Jellyfish from this package. I changed the Meterbridge properties become: /usr/bin/meterbridge -t jf alsa_pcm:playback_1 alsa_pcm:playback_2 My problem come here, i can open the Jellyfish window but i can't show the wave from input jack. How should i do? have you ever try this? some tell me to set up the Jack Audio Connection Kit, But i don't understand how to do it because i'm new for this

    Read the article

  • How to fix these compiler errors?

    - by Sandra Schlichting
    I have this source code from 2001 that I would like to compile. It gives this: $ make g++ -O99 -Wall -DLINUX -pedantic -c -o audio.o audio.cpp In file included from audio.cpp:7: audio.h:14: error: use of enum ‘mad_flow’ without previous declaration audio.h:15: error: use of enum ‘mad_flow’ without previous declaration audio.h:17: error: use of enum ‘mad_flow’ without previous declaration audio.cpp: In function ‘mad_flow audio::input(void*, mad_stream*)’: audio.cpp:19: error: new declaration ‘mad_flow audio::input(void*, mad_stream*)’ audio.h:14: error: ambiguates old declaration ‘int audio::input(void*, mad_stream*)’ audio.h:11: error: ‘size_t audio::stream::BufferPos’ is private audio.cpp:23: error: within this context audio.h:11: error: ‘size_t audio::stream::BufferSize’ is private audio.cpp:23: error: within this context audio.h:10: error: ‘char* audio::stream::Buffer’ is private audio.cpp:26: error: within this context audio.h:11: error: ‘size_t audio::stream::BufferSize’ is private audio.cpp:26: error: within this context audio.h:11: error: ‘size_t audio::stream::BufferPos’ is private audio.cpp:27: error: within this context audio.h:11: error: ‘size_t audio::stream::BufferSize’ is private audio.cpp:27: error: within this context audio.cpp: In function ‘mad_flow audio::output(void*, const mad_header*, mad_pcm*)’: audio.cpp:49: error: new declaration ‘mad_flow audio::output(void*, const mad_header*, mad_pcm*)’ audio.h:15: error: ambiguates old declaration ‘int audio::output(void*, const mad_header*, mad_pcm*)’ audio.cpp: In function ‘mad_flow audio::error(void*, mad_stream*, mad_frame*)’: audio.cpp:83: error: new declaration ‘mad_flow audio::error(void*, mad_stream*, mad_frame*)’ audio.h:17: error: ambiguates old declaration ‘int audio::error(void*, mad_stream*, mad_frame*)’ audio.cpp: In constructor ‘audio::stream::stream(const char*)’: audio.cpp:119: error: ‘input’ was not declared in this scope audio.cpp:122: error: ‘output’ was not declared in this scope audio.cpp:123: error: ‘error’ was not declared in this scope make: *** [audio.o] Error 1 audio.h contains #include <stdlib.h> #include "mad.h" namespace audio { class stream { private: char* Buffer; size_t BufferSize, BufferPos; struct mad_decoder Decoder; friend enum mad_flow input(void* Data, struct mad_stream* MadStream); friend enum mad_flow output(void* Data, const struct mad_header* Header, struct mad_pcm* PCM); friend enum mad_flow error(void* Data, struct mad_stream* MadStream, struct mad_frame* Frame); public: stream(const char* FileName); ~stream(); void play(); }; } I have tried to just insert enum mad_flow {}; but that just gave a new problem. Can anyone see how to fix this?

    Read the article

  • laptop headphone jack problems

    - by Xitcod13
    A while back my headphones mysteriously started making static noise and one of them stopped working completely. At first I thought it was headphones so I bought new ones. Alas that did not solve the problem. The problem must be inside my headphone jack. I did some research online and they suggested unplugging USB devices. Which has a strange effect of changing the static noises to high frequency Morse code noises (it's the aliens). I don't have this problem when i listen to music on speakers. The static is there on headphones whether there is music or not. I own a soldering iron for electronics and I am quite skilled at soldering. I would appreciate any help I can get. My laptop is the HDX 18. It has 2 headphone jacks that act exactly the same. Interesting thing i just noticed is that when i pull out my headphones almost all the way both of them start working but so do the speakers making the headphones kinda useless. Maybe there is a way to turn of the speakers as a temporary solution. I am using vista x64.

    Read the article

  • Is there an audio recording application/tool that has Tivo-like functionality?

    - by Bob
    I do a lot of live speech recording that requires me to quickly jump back and then transcribe a particular piece of the audio, then go back to recording again, while still maintaining the full audio file. So Far I've done this by splitting the audio and running one line to a recorder (for the whole audio), and one to my computer. Then I use something like Audacity to record, and then stop/go back whenever I hear something worth transcribing. This requires me to stop the recording, then start it up again and I end up missing chunks of the speech I'm listening to. Is there a tool that would let me rewind, then listen again and continue listening at a buffered distance from the audio recording, the way Tivo does with television shows?

    Read the article

  • HTML Audio performance

    - by user1888309
    I'm working on HTML drum machine, and I`ve met some performance issues, rhythm start to break if BPM is higher than 110 but I'm expecting to make it work on BPM over 180. I guess that it can be related with format or codec of audio files, however it also maybe that my code is not very optimised (as I can see from JS CPU profiling it's not). So I'm expecting you guys give me some code review or some hints on optimisation. Although all similar projects I've found on internet didn't work good and maybe it's just restrictions of Audio API. By the way, it's very raw and sounds works only on Chrome under Mac OS, so any advise on audio encoding for web also would be great Project on Github pages Screenshot of Groove which breaks UPDATE Ok, I've found that I was encoding audio files incorrectly, after fixing that rhythm stopped breaking, and also it started working in Mozilla. But still there are issues on windows OS.

    Read the article

  • iPhone 3.5mm jack based application

    - by maverick
    I want to encode data via a DTMF encoder and send it back to the iPhone via the 3.5mm Jack. Is it possible to send data back into the 3.5mm jack. conventionally audio signals are sent out over the iPhone 3.5mm jack? Is there provision to deal with DTMF and 3.5mm jack based input applications in Iphone's External Accessory framework?

    Read the article

  • Multiple audio sources on a single gameObject in unity

    - by angryInsomniac
    So, I have an audio system set up wherein I have loaded all my audio clips centrally and play them on demand by passing the requesting audioSource into the sound manager. However, there is a complication wherein if I want to overlay multiple looping sounds, I need to have multiple audio sources on an object, which is fine , so I created two in my script instantiated them and played my clips on them and then the world went crazy. For some reason, when I create two audio Sources in an object only the latest one is ever used, even if I explicitly keep objects separated, playing a clip on one or the other plays the clip on the last one that was created, furthermore, either this last one is not created in the right place or somehow messes with the rolloff rules because I can hear it all across my level, havign just one source works fine, but putting a second one on it causes shit to go batshit insane. Does anyone know the reason / solution for this ? Some pseudocode : guardSoundsSource = (AudioSource)gameObject.AddComponent("AudioSource"); guardSoundsSource.name = "Guard_Sounds_source"; // Setup this source guardThrusterSource = (AudioSource)gameObject.AddComponent("AudioSource"); guardThrusterSource.name = "Guard_Thruster_Source"; // setup this source // play using custom Sound manager soundMan.soundMgr.playOnSource(guardSoundsSource,"Guard_Idle_loop" ,true,GameManager.Manager.PlayerType); // this method prints out the name of the source the sound was to be played on and it always shows "Guard_Thruster_Source" even on the "Guard_Idle_loop" even though I clearly told it to use "Guard_Sounds_source"

    Read the article

  • Play an audio file using RemoteIO and Audio Unit

    - by NeilMonday
    I am looking at Apple's 'aurioTouch' example for the iPhone and I would like to play an mp3 or wav instead of using the built in mic. I am very new to the audio portion of iPhone programming, but I think I need to modify the SetupRemoteIO(...) function and replace the AudioComponent named 'comp' with a custom AudioComponent that plays a file. Basically I want the app to function exactly the same as the original, but with an audio file as the input instead of the mic.

    Read the article

  • Audio decoding delay when changing the audio language

    - by mahendiran.b
    My gstreamer Pipeline is like this Approach1 --------------input-selector->Queue->AduioParser->AudioSink | Souphttpsrc->tsdemux-->| | --------------- Queue->videoParser->videoSink In this approach 1, there is a delay in audio decoding when I toggle between various audio language. Approach2 ------ input-selector-> Queue->AduioParser->AudioSink | Souphttpsrc->tsdemux---multiqueue>| | ------- Queue->videoParser->VideoSink But there is no delay is observed in approach2. Can anyone please explain the reason behind this ? what is the specialty of multiqueue here?

    Read the article

  • Mini-jack problem with Sony Vaio (running XP)

    - by qftme
    I have a five year old Sony Vaio laptop (vgn-fw31m) that has had impact damage to the audio-output mini-jack for about the last year or so. In a recent discussion with my brother, we wondered whether it would be possible to write a program that would enable windows to use the microphone mini-jack input as the audio-output? As I currently use this laptop for work I am not keen to risk pulling it apart in order to replace the components comprising the audio-out. I therefore 'hope' that a programming solution exists. I would really appreciate any advice on this and eagerly await your response. Kind regards, qftme :)

    Read the article

  • Embed audio broadcasting on web page

    - by giargo
    Hi, I'd like to embed simple audio player on my webpage and I want it to get the audio from a stream broadcasted from my server. I read I can use IceCast on my web-server, getting an audio stream from a client using IceS (or this is what i got from other questions and articles) but once I have my stream, IceCast is supposed to broadcast it on an URL, that can be opened from pkayers like winamp or similar. I've found out this is quite a rare topic, usually people just want to broadcast "radio" where files are taken from a static playlist. In this case I have to get a stream from an IceCast URL and embed it with a player on a web page. Thank.

    Read the article

  • Audio libraries for PC indie games [closed]

    - by bluescrn
    Possible Duplicate: Cross-Platform Audio API Suggestions What options are out there these days for audio playback/mixing in C++? Primarily for Windows, but portability (particularly to Mac and iOS) would be desirable. For a small indie game, potentially commercial, though - so I'm looking for something free/low-cost. My requirements are fairly basic - I don't need 3D sound, or many-channels - simple stereo is fine. Just need to be able to mix sound effects and a music stream, maybe decoding one or more compressed audio formats (.ogg/.mp3 etc), with all the basic controls over looping, pitch, volume, etc. Is OpenAL more-or-less the standard choice, or are there other good options out there?

    Read the article

  • Verizon SongID - How is it programmed?

    - by CheeseConQueso
    For anyone not familiar with Verizon's SongID program, it is a free application downloadable through Verizon's VCast network. It listens to a song for 10 seconds at any point during the song and then sends this data to some all-knowing algorithmic beast that chews it up and sends you back all the ID3 tags (artist, album, song, etc...) The first two parts and last part are straightforward, but what goes on during the processing after the recorded sound is sent? I figure it must take the sound file (what format?), parse it (how? with what?) for some key identifiers (what are these? regular attributes of wave functions? phase/shift/amplitude/etc), and check it against a database. Everything I find online about how this works is something generic like what I typed above. From audiotag.info This service is based on a sophisticated audio recognition algorithm combining advanced audio fingerprinting technology and a large songs' database. When you upload an audio file, it is being analyzed by an audio engine. During the analysis its audio “fingerprint” is extracted and identified by comparing it to the music database. At the completion of this recognition process, information about songs with their matching probabilities are displayed on screen.

    Read the article

  • How to get musicbrainz track information from audio file

    - by Baki
    Can anyone tell me how to get track information from the MusicBrainz database from an audio file (mp3, wav, wma, ogg, etc...) using audio fingerprinting. I'm using MusicBrainz Sharp library, but any other library is ok. I've seen that you must use the libofa library, that you can't use MusicBrainz Sharp to get puid from the audio file, but I can't figure out how to use libofa with C#. Please show some examples and code snippets to help me, because I can't find them anywhere. Thanks in advance!

    Read the article

  • Debian sound on hdmi instead of jack

    - by Hans de Jong
    I installed debian (gnome) and i can't get my sound working. When i use inxi -A i get the following result: Audio: Card-1: Advanced Micro Devices [AMD] nee ATI Cayman/Antilles HDMI Audio [Radeon HD 6900 Series] driver: snd_hda_intel Card-2: Advanced Micro Devices [AMD] nee ATI SBx00 Azalia (Intel HDA) driver: snd_hda_intel Sound: Advanced Linux Sound Architecture ver: 1.0.24 My feeling tells me my sound output is on the HDMI instead of my jackplug on my motherboard. How can i change this to my motherboard sound output?

    Read the article

  • What are my audio devices?

    - by hellocatfood
    I'm attempting to use easycap to record from my camcorder but I'm having a slight problem. Using their test script I'm able to get audio and video. I've noticed that in the script on line 159 it makes a call to "DEV_ADUIO", which is reported as being "plughw:2,0". Exactly what is this device? Is it located in /dev/ somewhere? I've done "ls /dev/" and I can't find anything that would suggest an audio device

    Read the article

  • How to find an audio file's length (in seconds)

    - by mIL3S
    Hi all! (Objective C) Just using simple AudioServicesPlaySystemSoundID and its counterparts, but I can't find in the documentation if there is already a way to find the length of an audio file. I know there is AudioServicesGetPropertyInfo, but that seems to return a byte-buffer - do audio files embed their length in themselves and I can just extract it with this? Or is there perhaps a formula based on bit-rate * fileSize to convert to length-of-time? mIL3S www.milkdrinkingcow.com

    Read the article

  • Audio not working

    - by user3215
    Anybody could help me in troubleshooting audio problem on ubutnu 9.04 desktop edition?. For some reason I've to keep this os not upgraded and I'm trying to fix the audio problem on this for months. It works well on upgraded version(9.10,10.04) but not on jaunty. aplay -l: **** List of PLAYBACK Hardware Devices **** card 0: Intel [HDA Intel], device 0: ALC883 Analog [ALC883 Analog] Subdevices: 0/1 Subdevice #0: subdevice #0 card 0: Intel [HDA Intel], device 1: ALC883 Digital [ALC883 Digital] Subdevices: 1/1 Subdevice #0: subdevice #0 lsmod | grep snd: snd_hda_intel 436148 7 snd_pcm_oss 46336 0 snd_mixer_oss 22656 1 snd_pcm_oss snd_pcm 83076 4 snd_hda_intel,snd_pcm_oss snd_seq_dummy 10756 0 snd_seq_oss 37760 0 snd_seq_midi 14336 0 snd_rawmidi 29696 1 snd_seq_midi snd_seq_midi_event 15104 2 snd_seq_oss,snd_seq_midi snd_seq 56880 6 snd_seq_dummy,snd_seq_oss,snd_seq_midi,snd_seq_midi_event snd_timer 29704 2 snd_pcm,snd_seq snd_seq_device 14988 5 snd_seq_dummy,snd_seq_oss,snd_seq_midi,snd_rawmidi,snd_seq snd 62756 21 snd_hda_intel,snd_pcm_oss,snd_mixer_oss,snd_pcm,snd_seq_oss,snd_rawmidi,snd_seq,snd_timer,snd_seq_device soundcore 15200 1 snd snd_page_alloc 16904 2 snd_hda_intel,snd_pcm cat /proc/asound/cards: 0 [Intel ]: HDA-Intel - HDA Intel HDA Intel at 0xe1280000 irq 16 cat /proc/asound/version: Advanced Linux Sound Architecture Driver Version 1.0.18rc3. vim /etc/modules: # /etc/modules: kernel modules to load at boot time. # # This file contains the names of kernel modules that should be loaded # at boot time, one per line. Lines beginning with "#" are ignored. lp Audio Settings:

    Read the article

  • SEHException throw using Microsoft XACT Audio Framework (XACT3)

    - by Sweta Dwivedi
    I have been developing a game using Kinect + XNA and using Microsoft Audio Creation tool (XACT3) for managing my sound files and music, however in the code an SEHException is thrown whenever it tries to get the wave file from the wave Bank . . Sometimes the code works magically and all of a sudden it will start throwing this exception randomly ..I need a help on solving this exception /*Declaring Audio Engine for music*/ AudioEngine engine; SoundBank soundBank; WaveBank waveBank; Cue cue; /*Declaring Audio engine for sound effects*/ AudioEngine engine1; SoundBank soundbank; WaveBank wavebank; Cue effect; engine = new AudioEngine(@"Content\therapy.xgs"); soundBank = new SoundBank(engine, @"Content\Sound Bank.xsb"); **waveBank = new WaveBank(engine, @"Content\Wave Bank.xwb");** cue = null; engine1 = new AudioEngine(@"Content\Music_Manager\Sound_effects.xgs"); soundbank = new SoundBank(engine1, @"Content\Music_Manager\Sound1.xsb"); **wavebank = new WaveBank(engine1, @"Content\Music_Manager\Wave1.xwb");** effect = null; cue = soundBank.GetCue("hypnotizing"); cue.Play();

    Read the article

  • Physics-based dynamic audio generation in games

    - by alexc
    I wonder if it is possible to generate audio dynamically without any (!) audio assets, using pure mathematics/physics and some input values like material properties and spatial distribution of content in scene space. What I have in mind is something like a scene, with concrete floor, wooden table and glass on it. Now let's assume force pushes the glass towards the edge of table and then the glass falls onto the floor and shatters. The near-realistic glass destruction itself would be possible using voxels and good physics engine, but what about the sound the glass makes while shattering? I believe there is a way to generate that sound, because physics of sound is fairly known these days, but how computationaly costy that would be? Consumer hardware or supercomputers? Do any of you know some good resources/videos of such an experiment?

    Read the article

  • Playing part of a sfx audio file in HTML5 using WebAudio

    - by Matthew James Davis
    I have compiled all of my sound effects into one sequenced .ogg file. I have the start and stop times for each sound effect. How do I play the individual effects? That is, how do I play part of an audio file. More specificially, I've created a dictionary { 'sword_hit': { src: 'sfx.ogg', start: 265, // ms length: 212 // ms } } that my play_sound() function can use to look up 'sword_hit' and play the correct audio file at the correct start time for the correct duration. I simply need to know how to tell the WebAudio API to start playing at start ms and only play for length ms.

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >