Search Results

Search found 6011 results on 241 pages for 'fast fish'.

Page 120/241 | < Previous Page | 116 117 118 119 120 121 122 123 124 125 126 127  | Next Page >

  • SSD Performance for PHP?

    - by Andrew Fashion
    My programmer just built an application with PHP using Doctrine ORM (will be a high traffic social networking website), and it's very heavy in PHP/Apache and CPU. The queries are wonderfully fast, and MySQL is barely using any CPU, it's just Apache. I was curious to if an SSD would help speed up PHP/Apache, because I know the bottleneck is in PHP reading multiple files, class files, and loading up a bunch of data. So common sense makes me think if PHP is reading multiple PHP files, an SSD would only help as far as read/write? I was thinking of doing a high performance SSD for the PHP application, but for user image uploads, I would just continue using a 15k SAS. Is there any performance issues regarding using an SSD in this kind of situation? And would it prove to help speed up PHP/Apache, and help the CPU problem out?

    Read the article

  • When to use MySQL replication or DRBD for HA on Xen VM?

    - by user62513
    I'm setting up a database which needs to be needs to provide High Availabilty. My primary concern is high performance and robustness (I don't want something that will fail fast and badly). The database is accessed by the application at an average of 300 qps. It's will run on Xen VMs and it has some InnoDB tables as well as MyISAM tables. The VMs are connected via ethernet 100Mbit/s ethernet cables. Which of the two - MySQL replication or DRBD - would you recommend in such a situation? Or should I use DRBD to make the master database Highly Available and use MySQL replication on the slaves? I'm a developer so these things are all not so easy for me to make a sound judgement.

    Read the article

  • How can I tell if my live web-server is overloaded?

    - by Nick G
    We have a live webserver which doesn't seem to be performing all that well. It's a Dell PowerEdge machine, a few years old (dual core, 4GB) which is hosting about 20 low-traffic websites. However it doesn't seem to be as fast as it used to be. How can we determine the cause of this? If it's website traffic, I would be expecting high CPU but CPU usage is quite low and hovers around the 15-30% mark except for very brief periods. I'm wondering perhaps, if rather than CPU performance being a problem, perhaps it's disk thrashing due to the constant read/writes of all the small web files and database queries. It has 4x 7200 RPM SATA drives in RAID 5. So is there a way to check that it's not disk thrashing?

    Read the article

  • Can I set Windows default second-monitor behaviour to "Extend these displays"?

    - by MT_Head
    I travel to multiple offices (and multiple desks in those offices), and whenever possible I plug an external monitor into my laptop. Whenever I plug in a monitor I haven't used before, Windows defaults to "Duplicate these displays" - which messes up the arrangement of icons on my desktop if the external monitor is a different shape from my laptop's monitor. I then select "Extend these displays", and my laptop screen returns to its original shape - but my icons don't go back to their original arrangement. Grrrrr. Fast-forward a few days or weeks; I've got my icons arranged so I can find stuff again - then I go to a new office and it starts all over again. I'm tired of this. Is it possible to make "Extend these displays" the default behavior? I'm using Windows 8 x64 Home Premium, but I had the same complaint under Windows 7 x64 Ultimate. (Prior to that, I hadn't discovered the joy of dual displays. Ah, the time I wasted...)

    Read the article

  • Whats the best way to deal with backups for my php/mysql application

    - by spirytus
    I'm creating php application for my client and now thinking what would be the best way to do backups, automatically if possible? I don't have much experience in this area and in case something goes wrong, or if I need to migrate, I would like to have fast way of getting it all back online. I understand "something goes wrong" is a very wide term, but lets say that someone hacks my site and wipes out database and all the files. My app. is written in php/mysql and I got access to cpanel (hosted with justhost.com if that makes any difference :). I used Joomla and it has JoomlaPack that does complete backup almost automatically and in case site fails, its easy to revert, or migrate if necessary. Is there anything like that for my configuration that would make reverting/migration, easy?

    Read the article

  • I try to access a NFS mount via FTP. It works but the FTP Dir listing is very slow

    - by W0bble
    I mount an NFS using this command: mount -o rsize=8192,wsize=8192,timeo=14,intr serverip:/directory /mnt/directory However the mount appears on the client as expected a cmd like "ls -a" work pretty fast on the nfs mount. But when I try to list the mounted directory via FTP it gets very very slow ( 1.250 bytes in 160,39s (0,01KB/s) ). But surprisingly downloading files via FTP from nfs work with normal speeds. I tested several values for rsize and wsize parameter with no success. Both client and server are running Debian squeeze and NFSv4

    Read the article

  • Printing Booklet Page Size in Adobe Reader 4-in-1

    - by Justin Nathanael Waters
    So I have a 70 page pdf document that I'm trying to condense to a small booklet. I tried creating a formula to manually to perform it but it got ugly fast. 35,36,34,37,17,54,16,55,33,38,32,39,15,56,14,57,31,40,30,41,13,58,12,59 29,42,28,43,11,60,10,61,27,44,26,45,9,62,8,63,25,46,24,47,7,64,6,65 23,48,22,49,5,66,4,67,21,50,20,51,3,68,2,69,19,52,18,53,1,70 Once I print the booklet I should be able to cut the sheets in half and set the bottom half behind the top and staple it for a simple book. Which means Page 1 should have pages 35,36,17,54,34,37,16,55 Page 2 should have pages 33,38,15,56,32,39,14,57 And several pages later Page 9 should have pages 19,52,1,70,18,53 But manually doing this is a headache and it seems like the booklet function should contain functionality that can perform this. I'm using a commercial Konica Minolta C452

    Read the article

  • So confused by these CPU Specs can someone please help me out? THanks!

    - by Kevin
    Intel® Core™ i7-640M (2.8~3.46GHz, 35W) w/4MB Cache - 2 Cores, 4 Threads - 2.5 GT/s SO i'm buying a new laptop, which i have not done in 6 years. So i am not familiar with any of these cpu specs. It was the highest option for intel for this laptop. So i am assuming it is somewhat fast. But i'd like to learn what these specs mean. Any help would be greatly appreciated. i am not really a computer guy but would love to learn about what I am buying. Thanks!

    Read the article

  • Passing custom info to mongrel_rails start

    - by whaka
    One thing I really don't understand is how I can pass custom start-up options to a mongrel instance. I see that a common approach is the use environment variables, but in my environment this is not going to work because my rails application serves many different clients. Much code is shared between clients, but there are also many differences which I implement by subclassing controllers and views to overload or extend existing features or introduce new ones. To make this all work, I simply add the paths to client specific modules the module load path ($:). In order to start the application for a particular client, I could now use an environment variable like say, TARGET=AMAZONE. Unfortunately, on some systems I'm running multiple mongrel clusters, each cluster serving a different client. Some of these systems run under Windows and to start mongrel I installed mongrel_services. Clearly, this makes my environment variable unsuitable. Passing this extra bit of data to the application is proving to be a real challenge. For a start, mongrel_rails service_install will reject any [custom] command line parameters that aren't documented. I'm not too concerned as installing the services using the install program is trivial. However, even if I manage to install mongrel_services such that when run it passes the custom command line option --target to mongrel_rails start, I get an error because mongrel_rails doesn't recognize the switch. So here were the things I looked at: Pass an extra parameter: mongrel_rails start --target XYZ ... use a config file and add target:XYZ, then do: mongrel_rails start -C x:\myapp\myconfig.yml modify the file: Ruby\lib\ruby\gems\1.8\gems\mongrel-1.1.5-x86-mswin32-60\lib\mongrel\command.rb Perhaps I can use the --script option, but all docs that I found on it were for Unix 1 and 2 simply don't work. I played with 4 but never managed it to do anything. So I had no choice but to go with 3. While it is relatively simple, I hate changing ruby library code. Particularly disappointing is that 2 doesn't work. I mean what is so unreasonable about adding other [custom] options in the config file? Actually I think this is a fundamental piece that is missing in rails. Somehow, the application should be able to register and access command line arguments it expects. If anybody has a good idea how to do this more elegantly using the current infrastructure, I have a chocolate fish to give away!!!

    Read the article

  • Overheating on Ubuntu 12.04

    - by mati
    I have a Dell Inspiron Q17R with two graphic cards and I noticed that it is overheating. I installed Bumblebee, Jupiter and Flashblock, and I followed this guide as well, but it still got up to 74C. Is there anything more I can do? It still doesn't really seem to be working well and the fan keeps spinning really fast. After performing the following test: sensors in the terminal, this is what I got: Adapter: Virtual device temp1: +75.0°C (crit = +100.0°C) temp2: +75.0°C (crit = +100.0°C) coretemp-isa-0000 Adapter: ISA adapter Physical id 0: +68.0°C (high = +86.0°C, crit = +100.0°C) Core 0: +68.0°C (high = +86.0°C, crit = +100.0°C) Core 1: +65.0°C (high = +86.0°C, crit = +100.0°C) it doesn't look good.

    Read the article

  • Configure redis to not have everything in memory?

    - by acidzombie24
    I like redis because it lets me do operations on data structures. I wanted to see what would happen if i were to put more data into redis then i have for ram. So i wrote a loop that inserted 30k bytes repeatedly and set maxmemory 100MB. I figure it would stay at 100mb. It kept growing. Past 1gb then past 2gb. Suddenly it crashed because i ran the 32bit version. Now... i dont understand what the point of maxmemory is? I am using the windows version so maybe its ignored. Does redis have to have everything in memory? If i have a site (on linux) with a 10gb database and 512mb on the machine will redis work? I don't need it to be amazingly fast i just prefer to modify data in it then sql (although i hope it is still faster then mysql)

    Read the article

  • Remotely running batches on a Windows PC

    - by Eduardo León
    I want to remotely control my home desktop PC (running Windows 7 Professional), mainly to perform the following tasks: Downloading email attachments, and sending emails with attachments Running UI-less programs whose only inputs are files and whose only outputs are files So far, the only solution I have found is to use Remote Desktop to connect to my PC, but this is very slow and inefficient, especially when there is no fast Internet connection available other than my cell phone's. I would like to be able to send batch commands to my PC, like: Download an email attachment Use it as input for an UI-less program Save the program's output to a file Send that file to myself as an email attachment Is this possible? How could I do it?

    Read the article

  • Mail Secure & Stable Open Source Mail Server

    - by Fanar ALHAYALI
    I have asked question on http://stackoverflow.com/questions/9868426/i-need-to-know-which-email-server-i-have-to-use and someone tell me my question would be better on serverfault. I know that this is a common question and asked many times. but there are so many available mail servers that i am not able to decide the one. Kindly tell that which is the Secure, Stable and fast open source mail server for Centos or Redhat Server. Is there any guide which can be used to deploy the mail server with all its components e.g. smtp, pop3, imap, spam, calender server, antivirus, DNS Setting. Currently I'm using sun messaging V6 which installed on Solaris 10 and my boss ask me to make a report for the best mail server today in the marketing? I tried to have a look on Google but I couldn't find interesting information for my report. Any advice would be appreciated.

    Read the article

  • How do I improve picture quality while streaming live football (soccer) from my Dell D600 to an HDTV?

    - by Bob
    I have fibre broadband with speeds up to 38mbs, my Dell D600 has its max 2gb ram and has an ATI Mobility RADEON 9000 4xAGP 32mb card in it...Its TV support it says is NTSC or PAL in S-video and composite modes with a 7-pin mini-DIN connector (optional S-video to composite video adapter cable) and a vga port which i am using at the moment... The laptop runs Windows XP, an 80g HD with only windows + necessary updates and anti virus software on it..... There is HDMI on the TV, but not the laptop Fairly slow moving and close up pictures arent too bad, but when the movment is fast(a shot on goal) or in the distance, I cant see the ball and the images go out of focus.

    Read the article

  • very Slow machine when searching the network

    - by Adam
    Hi We are using a Dell desktop machine with 2 GB RRAM & Pentium D 3.00 GHz. The machine is very slow when using our Sales program and searching over the local network. We have this problem on 2 machine which are the same spec - This one and another. All of our other machine (10) run fast with no issues. Our server is SBS 2008 and AV is AVG. All of our client machines are WinXp Pro SP3 using the latest drivers. Any ideas?

    Read the article

  • Windows 7 extremely long startup

    - by Tyler
    Windows 7 before it even gets to login. On the "starting windows" splash screen takes no less than 15 minutes to finish more like 25 minutes most times. During this time the hard drive activity led indicator is blinking maybe once every 20 seconds. When I finally get to the desktop everything runs normally. I have unplugged all peripherals with same result. Ideas? It's 32bit. 4gig memory. Fast CPU which I can't recall off the top of my head.

    Read the article

  • PHPMyAdmin running very slow over internet but fine locally

    - by columbo
    I connect to PHPMyAdmin remotely on a Centos server using my local PC via Firefox. Usually it's fine but today it's really slow (2 minutes to load a page), sometimes timing out. Other connections to the server are fine. The SSH command line is as fast as ever as is the GNOME dekstop over SSH. In fact on the GNOME desktop I can run PHPMyAdmin locally from its browser and it's as quick as ever (which is a solution to the problem of course). I've checked the various log files and seen nothing unusual, I've logged into the MySQL command line and the database is running fine without any slowing what so ever. So it just seems to be slow when I access PHPMyAdmin on the server from the browser on my remote PC (I've tried IE and Firefox, both are slow). Has anyone experienced this or have any ideas what the issue could be. Connecting via CLI through tunnel works OK - problem is in phpMyAdmin for sure. Cheers

    Read the article

  • Internet really slow or stops working on my computer but fine on my laptop and ipad

    - by Fr0zen1
    I know you are going to say that this has been answered already, but it hasnt. None of those solutions worked for me. So this is the problem, and it only occurs on my computer, my laptop,ipad and iphone are just fine. When I`m browsing the net, it laods webpages slow(my speed is 30mb down and 3mb up) and when I start downloading with a another application(steam, utorrent), my browser completely stops working, but the application thats downloading(steam, utorrent) continues to download with a fast spead(2-3mb/s) and I do not know why. I have this issue on all browsers. I have tried chrome, internet explorer and firefox. I have restored my windows but still the same issue and I also have kubuntu installed and it also has the same issue. I do not know why. Any help? Thank you

    Read the article

  • Moving files with batch files from one pc to a server, to a another pc - worried about disk corruption

    - by AnchientAnt
    I use scheduled tasks that calls a batch file, that calls more batch files to move about three files from a pc, to a server, then to multiple other pcs. It all happens very quickly, as they are small files. Are there any pitfalls for how fast these transfers happen? I'm just mildly concerned about causing some disk corruption somehow. I use logic like 1. Call MapToPc if files exist then move file to folder on server. Disconnect 2. Call SendtoPCs If files exist (the files just moved to the server) then MapToPCs Move all files Disconnect All of this happens in about 2 secs or less. edit: this on windows 7, server 2003, xp respectively

    Read the article

  • ZFS on top of iSCSI

    - by Solipsism
    I'm planning on building out a file server using ZFS and BSD, and I was hoping to make it more expandable by attaching drives stored in other machines in the same rack via iSCSI (e.g., one machine is running ZFS, and others have iSCSI targets available to be connected to by the ZFS box and added to zpools). Looking for other people who have tried this has pretty much lead me to resources about exposing iSCSI shares on top of ZFS, but nothing about the reverse. Primarily I have the following questions: Is iSCSI over gigabit ethernet fast enough for this purpose, or would I have to switch to 10GbE to get decent performance? What would happen when one of the machines running iSCSI targets disconnects from the network? Is there a better way to do this that I just am not clever enough to have realized? Thanks for any help.

    Read the article

  • Add users in Windows machine without AD

    - by guillem
    I have several development machines where I am the administrator. We are using AD in my organization but is maintained by and offshore IT group any request takes a long time. We are currently granting access to developers on development machines manually so it's a bit annoying to maintain although at least it's fast. We have also a lot of external consultants that need to use those machines for some time. Is there any tool or method to maintain a set of users synced on those machines without the need add them to an AD group?

    Read the article

  • How to set/keep directory permissions?

    - by Dylan
    I'm using CwRsync to connect from my Windows development machine to a linux webserver : rsync -avuz -e ./ssh --exclude=".svn" /cygdrive/c/xampp/htdocs/project123/ [email protected]:/home/user123/public_html This syncs my development project directory nicely and fast to the server. But after doing this, all directory properties are reset to the local user user123 only, so the website is not available anymore. I need to manually reset those properties. Why is this happening, and how to prevent it? PS. coming from a Windows environment I'm having a really hard time understanding rsync. I copied the above command from some examples... just need to get this one small thing working too...

    Read the article

  • Is there any trick to join and use Windows 8/8.1 with Samba 4 (4.1.6)?

    - by tenshimsm
    It seems that Samba doesn't like at all. I've followed various tutorials and I can't get Windows 8 to work properly with a Ubuntu Server as domain controller. This week i've downloaded ubuntu 14.04 lts and set a fast domain configuration. As usual all other Windows version (XP and 7) work but the newest M$ nightmare doesn't. In this try it doesn't even join the domain, keeps saying the my username or password are wrong. My /etc/samba/smb.conf # Global parameters [global] workgroup = DOMAIN realm = DOMAIN.LAN netbios name = DOM server role = active directory domain controller dns forwarder = 8.8.8.8 idmap_ldb:use rfc2307 = yes [netlogon] path = /var/lib/samba/sysvol/domain.lan/scripts read only = No [sysvol] path = /var/lib/samba/sysvol read only = No [test] directory mode = 0750 path = /SHARES/test read only = no Does anyone have a tutorial that really works? Because I've tried many, each one with different configurations that works only with the people that made them. And is there a way to import my old AD users, computers and ID in a way that I won't need to rejoin all computers?

    Read the article

  • Normalizing a table

    - by Alex
    I have a legacy table, which I can't change. The values in it can be modified from legacy application (application also can't be changed). Due to a lot of access to the table from new application (new requirement), I'd like to create a temporary table, which would hopefully speed up the queries. The actual requirement, is to calculate number of business days from X to Y. For example, give me all business days from Jan 1'st 2001 until Dec 24'th 2004. The table is used to mark which days are off, as different companies may have different days off - it isn't just Saturday + Sunday) The temporary table would be created from a .NET program, each time user enters the screen for this query (user may run query multiple times, with different values, table is created once), so I'd like it to be as fast as possible. Approach below runs in under a second, but I only tested it with a small dataset, and still it takes probably close to half a second, which isn't great for UI - even though it's just the overhead for first query. The legacy table looks like this: CREATE TABLE [business_days]( [country_code] [char](3) , [state_code] [varchar](4) , [calendar_year] [int] , [calendar_month] [varchar](31) , [calendar_month2] [varchar](31) , [calendar_month3] [varchar](31) , [calendar_month4] [varchar](31) , [calendar_month5] [varchar](31) , [calendar_month6] [varchar](31) , [calendar_month7] [varchar](31) , [calendar_month8] [varchar](31) , [calendar_month9] [varchar](31) , [calendar_month10] [varchar](31) , [calendar_month11] [varchar](31) , [calendar_month12] [varchar](31) , misc. ) Each month has 31 characters, and any day off (Saturday + Sunday + holiday) is marked with X. Each half day is marked with an 'H'. For example, if a month starts on a Thursday, than it will look like (Thursday+Friday workdays, Saturday+Sunday marked with X): ' XX XX ..' I'd like the new table to look like so: create table #Temp (country varchar(3), state varchar(4), date datetime, hours int) And I'd like to only have rows for days which are off (marked with X or H from previous query) What I ended up doing, so far is this: Create a temporary-intermediate table, that looks like this: create table #Temp_2 (country_code varchar(3), state_code varchar(4), calendar_year int, calendar_month varchar(31), month_code int) To populate it, I have a union which basically unions calendar_month, calendar_month2, calendar_month3, etc. Than I have a loop which loops through all the rows in #Temp_2, after each row is processed, it is removed from #Temp_2. To process the row there is a loop from 1 to 31, and substring(calendar_month, counter, 1) is checked for either X or H, in which case there is an insert into #Temp table. [edit added code] Declare @country_code char(3) Declare @state_code varchar(4) Declare @calendar_year int Declare @calendar_month varchar(31) Declare @month_code int Declare @calendar_date datetime Declare @day_code int WHILE EXISTS(SELECT * From #Temp_2) -- where processed = 0) BEGIN Select Top 1 @country_code = t2.country_code, @state_code = t2.state_code, @calendar_year = t2.calendar_year, @calendar_month = t2.calendar_month, @month_code = t2.month_code From #Temp_2 t2 -- where processed = 0 set @day_code = 1 while @day_code <= 31 begin if substring(@calendar_month, @day_code, 1) = 'X' begin set @calendar_date = convert(datetime, (cast(@month_code as varchar) + '/' + cast(@day_code as varchar) + '/' + cast(@calendar_year as varchar))) insert into #Temp (country, state, date, hours) values (@country_code, @state_code, @calendar_date, 8) end if substring(@calendar_month, @day_code, 1) = 'H' begin set @calendar_date = convert(datetime, (cast(@month_code as varchar) + '/' + cast(@day_code as varchar) + '/' + cast(@calendar_year as varchar))) insert into #Temp (country, state, date, hours) values (@country_code, @state_code, @calendar_date, 4) end set @day_code = @day_code + 1 end delete from #Temp_2 where @country_code = country_code AND @state_code = state_code AND @calendar_year = calendar_year AND @calendar_month = calendar_month AND @month_code = month_code --update #Temp_2 set processed = 1 where @country_code = country_code AND @state_code = state_code AND @calendar_year = calendar_year AND @calendar_month = calendar_month AND @month_code = month_code END I am not an expert in SQL, so I'd like to get some input on my approach, and maybe even a much better approach suggestion. After having the temp table, I'm planning to do (dates would be coming from a table): select cast(convert(datetime, ('01/31/2012'), 101) -convert(datetime, ('01/17/2012'), 101) as int) - ((select sum(hours) from #Temp where date between convert(datetime, ('01/17/2012'), 101) and convert(datetime, ('01/31/2012'), 101)) / 8) Besides the solution of normalizing the table, the other solution I implemented for now, is a function which does all this logic of getting the business days by scanning the current table. It runs pretty fast, but I'm hesitant to call a function, if I can instead add a simpler query to get result. (I'm currently trying this on MSSQL, but I would need to do same for Sybase ASE and Oracle)

    Read the article

  • Data transfer is extrem slow after partitioning extern usb drive

    - by user125912
    I bought an extern usb 3.0 drive with 500 gb capacity. OS is Windows 7. I use it with an usb 2.0 slot, no prob. Initially I used it without making several partitions and it was fast as hell. Then I had the great idea to make partitions, one for programs, one for data and one for backup. I chose the free EASEUS Partition Master 9.1.1. and ended up with these partitions: F:Apps, primary, NTFS, 100gb H:Data, logic, NTFS, 250gb B:Backup, logic, NTFS, 150gb THE PROBLEM: When I copy files from C: to F: I get a transfer rate of about 100 KB/S ! When I copy files from C: to H: I get a transfer rate of about 4 MB/S ! thats all muuuch to slow, slower then before. What can I do to speed the shit up? Thanks in advance!

    Read the article

< Previous Page | 116 117 118 119 120 121 122 123 124 125 126 127  | Next Page >