Search Results

Search found 11409 results on 457 pages for 'large teams'.

Page 224/457 | < Previous Page | 220 221 222 223 224 225 226 227 228 229 230 231  | Next Page >

  • 421 Concurrent Connections - Ratelimit from helpdesk to rackspace server

    - by g18c
    We have Kayako helpdesk running on our WHM Linux server. When e-mails come in from customers, notifications are sent out by Kayako to a number of staff whose mailboxes are hosted on Rackspace mail servers. I noticed a large queue in the Exim queued message viewer of WHM - when looking in Exim logs I can see many lines 2012-10-13 20:06:56 1TN72s-0007Cw-1l SMTP error from remote mail server after initial connection: host mx2.emailsrvr.com [173.203.2.32]: 421 Too many concurrent connections from this client. One client email results in about 5 emails to rackspace servers, perhaps 60 emails per 1 hour on average - not a huge amount but enough to cause messages to be rejected when sent in short bursts. In this case ideally if we can limit the connections sent to the rackspace server we can comply with their limit. For our requirements if we send 1 email every10 seconds or so, this would be OK. Messages to all other servers should go through a normal rates, only mx1.emailsrvr.com and mx2.emailsrvr.com should have this connection limit policy applied. Is this possible?

    Read the article

  • How can I suppress /etc/issue without losing error messages?

    - by Andy
    Is it possible to tell the ssh client to not print the connects of /etc/issue to stdout when connecting to a remote host, but to print out any other diagnostic (e.g. error) messages? Either using ssh -q or having LogLevel quiet in ~/.ssh/config suppresses the /etc/issue printing, but also turn off error messages. I've tried touching ~/.hushlogin as well - that stops /etc/motd being printed, but doesn't affect /etc/issue. The most obvious solution is just to remove /etc/issue, but company policy dictates the file be there with dire warnings about unauthorised access. This is non-negotiable. Unfortunately, I've got a bunch of scripts that run across quite a few hosts via ssh, and the log files are a) very large and b) full of legalese. Since quite a lot of stuff runs unattended, I don't want to lose any error messages that are printed.

    Read the article

  • converting huge MPEG audio files to something smaller

    - by john
    I've got some large MPEG audio files (144 MB each) that I'm looking to convert to something smaller so I can send them out as attachments to an email. Any suggestions on the software to use? I'm looking for something free that will run on Windows. I don't really care what the destination file is, mp3 would be nice. If there's a web service out there that would do this without the need to download any software to my machine, that would be even better, but I would be more than happy just getting it done any way I can. Thanks!

    Read the article

  • ML 350 additional SATA RAID controller (mirror only)

    - by Nicholas
    I have a Proliant ML350 G8 with two SAS raid arrays currently set up - thereby maxing the default P420i raid controller. I need to set up a large video dump space in addition to this existing set-up (for non backed up, non-critical, temporary storage). I had planned to just add a 2TB SATA disk and plug it into the motherboard. However I it occurred to me the motherboard might have built-in mirror RAID support? Therefore I could use two SATA disks and have some semblance of redundancy. Is this possible? Or would I need to get a cheap raid card? Any recommendations?

    Read the article

  • Combine Multiple Audio Files into a single higher-quality audio File

    - by namenlos
    BACKGROUND My team gave a demo to a large audience - we recorded the audio of the demo in multiple locations in the room (3) the audio was recorded using cheap laptop microphones I was not involved in the recording of the audio or the demo Both audio files suck in some form the first one is of a recording near the speaker - which clearly gets his voice but the the audience is audience is muffled - also this one is slightly noisy The second recording was done in the middle of the audience - it gets the audience questions clearly but actually gets the speaker rather sometimes well and sometimes poorly (not all the speakers spoke loudly enough to be heard) MY QUESTION Is there any techinque or software which can be used to merge these audio files in such a way that the best qualities of each are preserved. I am NOT asking now to simply merge them together in one track - I've already done that in Audacity and it is certainly better - what I am looking for could be considered closer to how HDR images are created - multiple exposures combined into an enhanced new version which is not simply an average of the inputs. NOTE Am not an "Audio" guy - just a normal user

    Read the article

  • Where does power consumption go in a computer?

    - by Johannes Rössel
    Today we had a weird discussion over lunch: What exactly causes power consumption in a computer, particularly in the CPU? Figures you usually see indicate that only a percentage (albeit a large one) of the power consumption ends up in heat. However, what exactly does happen with the rest? A CPU isn't (anymore) a device that mechanically moves parts, emits light or uses other ways of transforming energy. Conservation of energy dictates that all energy going in has to go out somewhere and for something like a CPU I seriously can't imagine that output being anything but heat. Us being computer science instead of electrical engineering students certainly didn't help in accurately answering the question.

    Read the article

  • pdf creation software, for docx and odt

    - by oxinabox.ucc.asn.au
    Ok, I have a fairly large collection of docx and odt files. Minutes from meetings etc. Now I want to convert them to pdfs for distrobution. and also into one combined pdf. At the momement I'm using Adobe Acrobat 8 (Pro iirc). and on another machine I'm using foxit pdf printer. To do this I have to print them each individually to pdfs. and then I can combine them with Acrobat, cos acrobat doesn't support conversion stright from docx or odt to pdf - only via printing. Now this is annoying if you have to do it on a regular basis, since i don't keep the pdfs around (I have the orignals source controlled :-D) cos they go out of date pretty quick as I often have to go back and modify old versions (like ridiculously often). e Eg When I find out I've got something in the minutes wrong or I want to add more context for clarifaction. Anyone got a better solution?

    Read the article

  • Windows Explorer view style via GPO or registry edit

    - by iamninja
    I'm having some difficulty adjusting the default view style in windows explorer (e.g. list, detail, large icon, small icon). There does not seem to be any pre-baked GPOs for this and next to no documentation on the subject outside of using GUI configuration tools built into windows explorer. Does anyone know the specific registry keys that need to be edited for this? I've already taken a look at HKCU\Software\Microsoft\Windows\CurrentVersion\Explorer\Streams\Settings but the adjustments I've made there don't seem to make any change on their own.

    Read the article

  • Cheapest server per gigabit throughput [closed]

    - by nethgirb
    I'm looking for a set of servers for performance testing a network, and secondarily testing some applications on the servers. Their most important task is simply to pump out data: from an application like memcached or just dumped from a large file in memory into a TCP flow (i.e., disk performance doesn't matter). This should happen over one or more 1 gigabit Ethernet ports, and the machines should run Linux (ideally), or perhaps Mac OS X or some other *nix. Other than that, there are few constraints (e.g., even something ARM-based could be fine). So here's the question: What's the cheapest server per gigabit? Price and power are both considerations.

    Read the article

  • How to track down a file descriptor leak?

    - by cclark
    I have a java process (Glassfish) which is leaking file descriptors. I know this because I get the helpful java.io.IOException: Too many open files exception. I can look in /proc/PID#/fd and see all the open file descriptors. When I use lsof I get a very large number of entries like this: java 18510 root 8811u sock 0,4 1576079 can't identify protocol java 18510 root 8812u sock 0,4 1576111 can't identify protocol java 18510 root 8813u sock 0,4 1576150 can't identify protocol I see 12 new ones created per minute. What options can I use on lsof or what other tools are available to me to help track down socket file descriptors where the protocol can't be identified? thanks, chuck

    Read the article

  • Why is my ogg bigger than my m4a?

    - by acidzombie24
    I am using the following commands to encode an audio file to m4a & ogg formats: ffmpeg.exe -i 0123456789 -ab 192k out.m4a ffmpeg.exe -i 0123456789 -f wav - | oggenc2.exe - -r -q 6 -o out.ogg (0123456789 has no extension.) My m4a output is 14,608kB while my ogg output is 19,809kB. Why? AFAIK -q 6 is roughly 192kbps. So it should be about even. I could see one file being 1-3MB bigger than the other, but 5MB is pretty large. The m4a is almost 75% of the ogg! Why is this?

    Read the article

  • EFI Partition Types

    - by Will
    Another employee was working with a mounted VHD and was using diskpart on it and I'm not sure if he messed something up or not. I think it's fine but I wanted to double check. We have a Windows 2008 x64 machine that boots with EFI because of the large disks in the system. What should the different labels be within Disk Manager for this system? Here is what is currently set: 200 MB Partition - EFI System Partition C Partition - Boot, Page File, Crash Dump There should not be an active partition under EFI, correct? Also, is there supposed to be a "System" label in addition to "EFI System Partition"?

    Read the article

  • Laptop and 2 screens: use screens but not monitor display

    - by ClarkeyBoy
    Hi, I have 1 VGA socket on my laptop, and currently have that in use by a large screen. At some point in the future I would like to get another one of these screens and use both screens in dual screen mode but not use my laptop display (to be honest my laptop display is pretty rubbish as its like 2/3s the size of my screen - even if I had the choice to use all 3 I probably wouldn't want to). Is it possible to achieve this? If so, what do I need by way of hardware / software, and how much do you reckon it should cost me? Thanks in advance. Regards, Richard

    Read the article

  • Extracting metadata from flac header in linux

    - by Tommy Fisk
    Hello. I recently (and accidentally) wiped a hard-drive which should NOT have been formatted. Thankfully, PhotoRec was able to recover some of the most important files from the disk. Part of those files include a large music collection of flac files. I've placed these in a folder named 'FLAC' that looks like this: [Start of FLAC Folder] f11655088.flac f11698672.flac ... (around 2,000 files total) f291142600.flac [End of FLAC Folder] I'm running a livecd to rescue this data. I have access to a standard terminal in Ubuntu 10.10. I've noticed that the music player in linux will display the song title when I open any of these songs. My goal is to automatically extract the song title from these files and update their names from f12313512.flac to 'Libera - Rebirth.flac' - or whatever the song title happens to be. Thanks in advance for suggestions on how I might accomplish this!

    Read the article

  • NFSv3 Asynchronous Write Depends on Block Size?

    - by Joe Swanson
    I am trying to figure out if my NFSv3 deployment is performing SAFE asynchronous writes. I suspect that it is doing strictly synchronous writes, as I am getting poor performance in general. I used Wireshark to look at the 'stable' flag in write calls, and look for 'commit' calls. I noticed that, with especially large block sizes, writes to appear to be performed asynchronously: dd if=/dev/zero of=/proj/re3/0/zero bs=2097152 count=512 However, smaller block sizes appear to be performed strictly synchronously: dd if=/dev/zero of=/proj/re3/0/zero bs=8192 count=655360 What gives? How does the client decide whether to tell the server to perform writes synchronously or asynchronously? Is there any way I can get smaller block sizes to be performed asynchronously?

    Read the article

  • Should I convert my AAC M4A files to MP3?

    - by j0rd4n
    Due to Apple, I have a large majority of my music files in the AAC M4A format. They do NOT have DRM so I don't have to worry about that. I'm getting tired of Apple products and really want to switch to a different brand player (and something more compatible with Linux). It appears most MP3 players support...well...MP3 and not AAC. Should I convert my library to be free of Apple and open to other players? Is this a lossless conversion? Can it be lossless? If I will lose quality, I'm not interested. Am I even doing the right thing? AAC is the better format, but I'm not seeing a lot of support for it yet. I'll be honest and say that I need some education in this department. Any helpful advice is most welcome.

    Read the article

  • Case for Micro SD card?

    - by Josh
    I have a MicroSD card which I'd like to keep with me at all times, I.E. in my wallet. I'm looking for a case for it. But all the cases I can find seem to be for standard SD cards... the closest I could find was this: Which is completly pointless, it stores the Micro SD card along with it's adaptor. Why not just put the Micro SD card IN the adaptor and store both on a normal SD card case... Anyway, does what I'm looking for (a case to protect a microSD card, and only large enough for a mircoSD card, i.e. not what's pictured above) exist?

    Read the article

  • What is the current state of development for the FLAC codec?

    - by TwentyMiles
    I have a pretty large collection of FLAC files created from my CD collection. I love the FLAC format and the sound quality that you can get from it. Lately, however, I've been trying to write a few tools to manipulate the files and I've been noticing what seems to be a stagnation of the community around the codec. Some of the links on the official FLAC page point to thing that are no longer relevant (7digital, for example, appears to no longer sell FLAC encoded songs). It's pretty hard to find hardware players that support FLAC any more (most noticeably it's not present on lower end players when it used to be, and playback is absent on Android). Programming language tools (Java and .NET libraries) are at best old, and at worst unfinished. What's the current state of FLAC development? Has it been replaced by another codec?

    Read the article

  • Can I split a spreadsheet into multiple files based on a column in Excel 2007?

    - by geofftnz
    Is there a way in Excel to split a large file into a series of smaller ones, based on the contents of a single column? eg: I have a file of sales data for all sales reps. I need to send them a file to make corrections and send back, but I dont want to send each of them the whole file (because I dont want them changing eachother's data). The file looks something like this: salesdata.xls RepName Customer ContactEmail Adam Cust1 [email protected] Adam Cust2 [email protected] Bob Cust3 [email protected] etc... out of this I need: salesdata_Adam.xls RepName Customer ContactEmail Adam Cust1 [email protected] Adam Cust2 [email protected] and salesdata_Bob.xls Bob Cust3 [email protected] Is there anything built-in to Excel 2007 to do this automatically, or should I break out the VBA?

    Read the article

  • Error 2020: Got packet bigger than 'max_allowed_packet' bytes when dumping table

    - by Imagineer
    I'm getting the above mentioned error when backing up with ZRM, which is using mysqldump for backup. mysqldump --opt --extended-insert --single-transaction --create-options --default-character-set=utf8 --user=" " -p --all-databases "/nfs/backup/mysql01/dailyrun/20091216043001/backup.sql" mysqldump: Error 2020: Got packet bigger than 'max_allowed_packet' bytes when dumping table TICKET_ATTACHMENT at row: 2286 I have increased the size for 'max_allowed_packet' to be 1G in /etc/my.cnf which is the server setting and for the client side setting I've set it by running this command: mysql -u -p --max_allowed_packet=1G And I have verified that on the client and server side they are of the same value. This is to check the client side value according to this forum posting http://forums.mysql.com/read.php?35,75794,261640 mysql SELECT @@MAX_ALLOWED_PACKET - ; +----------------------+ | @@MAX_ALLOWED_PACKET | +----------------------+ | 1073741824 | +----------------------+ 1 row in set (0.00 sec) And this is the check the server value setting. mysql SHOW VARIABLES | max_allowed_packet | 1073741824 | I have ran out of ideas, and tried searching within expert exchange and googling for solutions but so far none has worked. Reference http://dev.mysql.com/doc/refman/5.1/en/packet-too-large.html Anyone please advise, thank you.

    Read the article

  • Recommendation for e-mail forwarding

    - by Paul Mrozowski
    We currently have a provider that we use to primarily forward our e-mails to both a hosted Exchange site and to outside e-mail addresses. We have a large # of forwards set-up (maybe 500). They work fine for us but the bandwidth costs are a getting pretty expensive (approx 80GB in email a month) so we're looking for a less expensive alternative (it costs us nearly $150/mo). I've looked at just hosting this here via a PostFix server and one of the nice front-ends available for it. However, I'm not thrilled with the prospect of being responsible for maintaining yet another server, backups, network connectivity, etc. So I'm leaning towards having this hosted by someone else if possible. Does anyone have any recommendations for hosting providers? Are there some I should stay away from?

    Read the article

  • what software will rip a CD to flac files in one step?

    - by jcollum
    I'm using EAC to rip files to FLAC. Seems fine. But for a large amount of CDs (200 or so) the process is a little labor intensive. There's 4 or 5 clicks in there that seem extraneous -- EAC seems to want to rip the CD to a set of WAV files then (after all those files are ripped) I have to select "Process WAV files" (I'd get it exact, but EAC is occupied right now). I'd like to send all of it to FLAC files in just a few steps: get cddb info, select art, rip. Is there a way to do this with EAC that I'm missing? Haven't been able to find anything on the web that explains this. Is there a better program for this?

    Read the article

  • windows server 2008 vs ubuntu 11 [closed]

    - by user472875
    I am working on implementing a custom server application that should be capable of handling a very large volume of traffic. I am aware that this type of question has been asked a lot, but I haven't been able to find a good answer. What I'm really looking for is for a server with given specs which OS will be able to handle a larger traffic faster and more reliably. I do not care about rights management or any other features. I am fairly good with both platforms, and so I would like to pick the OS with better performance on a clean install, and with nothing else running. Thanks in advance.

    Read the article

  • IIS7 Custom ASP.NET Errors

    - by Nathan
    I'm trying to setup a custom error page for the IIS 7 404.13 (Content length too large) error. Here's the relevant sections of my web.config file: <system.webServer> <httpErrors errorMode="Custom" existingResponse="Replace"> <remove statusCode="404" subStatusCode="13" /> <error statusCode="404" subStatusCode="13" prefixLanguageFilePath="" path="/FileUpload/Test.aspx" responseMode="ExecuteURL" /> </httpErrors> <security> <requestFiltering> <requestLimits maxAllowedContentLength="10240" /> </requestFiltering> </security> </system.webServer> The response that is being sent back to the server is blank. The Test.aspx file is not blank. Any idea what's going on here?

    Read the article

  • how do applications (and OS) handle very big files?

    - by DrStrangeLove
    For instance, i have video file which is 11.8 Gb, but my RAM memory only 2 Gb.. How does VLC (or other software) handle it? How do they load it into memory? I used VMMap tool (from sysinternals) to take a look at memory, and i saw: private 160000K working set 100000K Obviously, it's much less than 11.8 Gb -So how did it happen? This question is not only about video. I'd like to know how computer, in general, handles very large files.

    Read the article

< Previous Page | 220 221 222 223 224 225 226 227 228 229 230 231  | Next Page >