Search Results

Search found 19279 results on 772 pages for 'everything'.

Page 37/772 | < Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >

  • RAID 0 performance gains?

    - by NickAldwin
    I'm building a new computer over the summer. I'm fairly competent in computer hardware, and am thus building the computer from scratch. I have everything planned out, but I was wondering about RAID. I asked which RAID I should use earlier, but now that it's pretty clear that RAID 1 isn't really that great, I think I'll go with cloud-backup instead of disk-redundancy. However, I still face a choice: use two 1TB drives as two 1TB drives, or combine them into a RAID 0 striped array. Is there any performance gain at all? I know that if one drive dies, everything is gone, so is the performance gain worth it? I'm building a pretty advanced computer, with SLI video cards and a fast CPU, so I'm thinking RAID 0 would give me some good hard drive performance. From your experience, is RAID 0 viable?

    Read the article

  • Win XP Pro SP3 MUP.SYS last driver to load before freeze. Can't boot into XP CD / Recovery Console

    - by Joshua
    I've tried everything and have even looked at the thread on this site "Fresh installation of XP hangs after MUP.SYS" but nothing. I'm running Memtest86+ v4.00 but no luck (everything looks good) It freezes at windows boot screen, not even 1 green block shows up in the loader. I go into safemode and it stops at MUP.SYS, it's probably not MUP.SYS that is the problem but it's probably the next file that is trying to load or something. The end-user has not installed anything at all that could be problematic. I have pulled the HDD out and put it into another pc and have checked it, nothing suspicious at all. It's just as clean as when I first cleaned it up and secured it. Someone / anyone please help! :) Thank you!

    Read the article

  • How to push changes from Test server to Live server?

    - by anonymous
    As a beginner, I finally noticed the issue with making changes to the live server I've been working on, now that I have a couple users on it, since I bring it down so often. I created an EC2 image of my live server and set up a separate instance on EC2, so now I have 2 EC2 instances, Stage and Production. I set up GitHub and push changes to stage and test my code there, and when it's all done and working, I push it to the production branch, and everything is good. And there is a slight issue here since I name my files config_stage.js and config_production.js and set up .gitignore on each server, and in my code, I would have it read the ENV flags and set up the appropriate configs, is this the correct approach? And my main question is: how do you keep track of non-code changes to the server? For example, I installed HAProxy, Stunnel, Redis, MongoDB and several other things onto the Stage server for testing and now that it's all working and good, how do I deploy them to production? Right now, I'm just keeping track of everything I installed and copying configuration files over, which is very tedious and I'm afraid I may have missed a step somewhere. Is there a better way to port these changes over from my test server to my live server?

    Read the article

  • BizTalk 2010 upgrade - Sunset Development/Deployment Modes

    - by Ahsan Alam
    Those who are familiar with BizTalk 2006, should know about Development and Deployment modes in Visual Studio. Personally, I never questioned why it's not Debug and Release just like everything else in Visual Studio. Then everything changed in BizTalk 2010. BizTalk and Visual Studio 2010 now uses Debug and Release modes by default. When we upgraded BizTalk 2006 solution to 2010, Development and Deployment modes remained unchanged for all the projects, and code compiled without any issues. Soon, I realized that any new projects added to the converted solutions started using Debug and Release modes. This also didn't cause any problem compiling the solution from Visual Studio; however, it broke our custom build/deployment scripts since the scripts were trying to build in Deployment mode. So, I decided to change all projects from Development and Deployment modes to Debug and Release modes to keep them consistent. During this process I realized that Debug and Release modes are defaults; but it's completely customizable. During the BizTalk 2010 upgrade process, I figured that switching to default Debug and Release modes are the best options.

    Read the article

  • Cisco 877 as a VPN server?

    - by Massimo
    I have a Cisco 877 router which connects my network to the Internet using an ADSL line, a single public IP address and NAT; the IOS version is 15. Everything is working ok, but I'd like to configure this router to be a VPN server, to be able to connect to the network from the outside. I've tried looking for documentation, but everything I can find is related to having the 877 act as a VPN client, or to site-to-site VPNs; I can't find anything about letting single remote computers access the internal network, which is something I can do quite easily using Windows' RRAS or ISA Server. Can the Cisco 877 act as a VPN server for remote client computers? (Looks like it should, but just to be sure...) Which type of VPNs does it support? Do they require some special software on the client machines, or can they be used by standard out-of-the-box Windows computers? And, finally: how to set up this?

    Read the article

  • A Firefox extension for scan & upload document?

    - by Ivan Petrushev
    Hello, Do you know such an extension that provides easy document scanning in Firefox? We are building a web site and we want visitors to be able to upload scanned documents to it. The normal procedure for that is: Scan the document via Gimp, Photoshop or some other scanning software. Save the file. Navigate to the upload web page. Find some sort of HTML input type file on that page. Browse and find the saved file. Submit the form. I want an extension or plugin that automatize that process and do everything with 1 click - scan the document with some default settings (for example "grayscale, 300 dpi") creates temporary file, fills the page input field and deletes the temporary file after upload. I tried lots of googling but the term scan in combination with everything web-related gives zillions of virus, malware and port scanners...

    Read the article

  • Custom Profile Provider with Web Deployment Project

    - by Ben Griswold
    I wrote about implementing a custom profile provider inside of your ASP.NET MVC application yesterday. If you haven’t read the article, don’t sweat it.  Most of the stuff I write is rubbish anyway. Since you have joined me today, though, I might as well offer up a little tip: you can run into trouble, like I did, if you enable your custom profile provider inside of an application which is deployed using a Web Deployment Project.  Everything will run great on your local machine and you’ll probably take an early lunch because you got the code running in no time flat and the build server is happy and all tests pass and, gosh, maybe you’ll just cut out early because it is Friday after all.  But then the first user hits the integration machine and, that’s right, yellow screen of death. Lucky you, just as you’re walking out the door, the user kindly sends the exception message and stack trace: Value cannot be null. Parameter name: type Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code. Stack Trace: [ArgumentNullException: Value cannot be null. Parameter name: type] System.Activator.CreateInstance(Type type, Boolean nonPublic) +2796915 System.Web.Profile.ProfileBase.CreateMyInstance(String username, Boolean isAuthenticated) +76 System.Web.Profile.ProfileBase.Create(String username, Boolean isAuthenticated) +312 User error?  Not this time. Damn! One hour later… you notice the harmless “Treat as library component (remove the App_Code.compiled file)” setting on the Output Assemblies Tab of your Web Deployment Project. You have no idea why, but you uncheck it.  You test and everything works great both locally and on the integration machine.  Application users think you’re the best and you’re still going to catch the last half hour of happy hour.  Happy Friday.

    Read the article

  • When SharePoint Designer has its own designs

    - by PeterBrunone
    Recently, a colleague came to me with a simple task and an inscrutable error.  He just wanted to populate a text field with a querystring value.  If you've ever done this in SPD, you know it's fairly simple:  create a parameter, map it to a querystring value, and then use the resulting parameter name in your form field. Having done so, however, he was told the following by the ASP.NET "yellow barf page": The 'Text' property of 'asp:TextBox' does not allow child objects. As it turns out, he had done everything correctly.  The problem was that SharePoint Designer had decided the best place for his FieldDescription control was INSIDE the TextBox control.  Obviously the compiler doesn't know what to do with that.  When the FieldDescription was moved to a less obtrusive location, everything worked as expected.The moral of the story is, as always, don't trust what any WYSIWYG tool gives you.  If it looks great, then fine.  However, if there's a problem, remember that Design mode was written by human beings who make mistakes... just like the rest of us.Take THAT, Skynet.

    Read the article

  • Virtualbox two networks slow

    - by Petr Marek
    I am running an Ubuntu server guest on Win 7 guest, and am running a webrick server (RoR dev). If I have just a host-only network, everything works fine and the browser response is instant. However, if I add a second network (NAT), so that the server can connect to the internet (for various updates etc.), the host-to-guest access gets really slow. I can't use the bridge connection. I am using the port 3000 (RoR Webrick server) and connecting to the guest via internet browser on this port (eg http://192.168.56.102:3000). Any idea, what could be causing this? If I ping the IP from host console, I get < 0ms. Here are the settings (relevant info is in english; Povoleno vše is Everything is allowed):

    Read the article

  • "SC.EXE config" and dollar sign in service name

    - by Joe Schmoe
    Say I want to make Windows Service startup dependent on SQL Server. In my case service name for SQL Server is MSSQL$SQL11 (SQL11 is SQL Server instance name) However, when I issue this command: SC.EXE config MyService depend= MSSQL$SQL11 everything after the dollar sign is ignored. When I go to "Dependencies" tab in "Services" SQL Server is not listed. When I check matching registry key it becomes clear why: it only has MSSQL. At this point I have to edit registry by hand to change MSSQL to MSSQL$SQL11 and then everything works as expected. Putting quotes around MSSQL$SQL11 doesn't help. Is there a way to specify $ in the middle of the SC.EXE argument string?

    Read the article

  • Rendering with Direct3D

    - by Jamie
    Hi, I'm slightly confused about how Direct3D rendering works. Basically, as long as I render to one surface, everything is fine. But when I try rendering to multiple surfaces, it seems like everything is still rendered to one surface. I think there's something wrong with my calls. For each update cycle this is what I do 1. device-BeginScene() 2. sprite-Begin(...) ... A bunch of GetRenderTarget to store the old render target, then SetRenderTarget to set a new surface, and then things like CreateVertexBuffer, SetTexture, etc to draw on the new render target. Then resetting to the old render target. sprite-Draw([the back buffer]) (the back buffer is actually another surface, not the actual back buffer. But here it is being drawn onto the actual back buffer, I think) sprite-End() device-EndScene() device-Present(...) Also, it seems like if I mix sprite drawing and non-sprite drawing onto a surface, that first one set of render commands is executed and then the other set, rather than in order by when each command was called. If anyone could shed light on any of this, it would be much appreciated.

    Read the article

  • Nvidia GeForce 410m

    - by user1789927
    I have a Lenovo v570 with a Nvidia GeForce 410m with cuda. It came with windows 7 and everything works well. However I just got bored of windows 7 and I'm yearning for xp. I installed xp and got everything to work except the Nvidia driver. In their website there is only the windows 7 nvidia driver available. I searched http://www.geforce.com/drivers/results/26388 and got a driver for the 400m series. I tried the software and it didn't work, message << THIS GRAPHICS DRIVER COULD NOT FIND COMPATIBLE GRAPHICS HARDWARE . I also tried : http://forums.laptopvideo2go.com/topic/29285-v29610-windows-xp-32bit-nvidia/ This time the error message 'REQUIRED FILES ARE MISSING' Any help will be grateful.

    Read the article

  • Where did my form go in SharePoint 2010!?

    - by MOSSLover
    So I was working on an intro to development demo for the Central NJ .Net User Group and I found a few kinks.  I opened up a form to custom in InfoPath and Quick Published it wouldn’t work.  I imed my InfoPath guru friend, Lori Gowin, she said try to run a regular publish.  The form was still not showing up in SharePoint.  I could open it and it knew my changes, but it would just not render in a browser.  So I decided to create a form from scratch without using the button customize form in the list.  That did not work, so it was google time.  Finally I found this blog post: http://qwertconsulting.wordpress.com/2009/12/22/list-form-from-infopath-2010-is-blank/.  Once I went into the configuration wizard and turned on the State Service everything worked perfectly fine.  It was great.  See I normally don’t run through the wizard and check the box to turn all the services on in SharePoint.  I usually like a leaner environment plus I want to learn how everything works.  So I guess most people had no idea what was going on in the background.  To get InfoPath to work you need the session state.  It’s doing some type of caching in the browser.  Very neat stuff.  I hope this helps one of you out there some day. Technorati Tags: InfoPath 2010,SharePoint 2010,State Service

    Read the article

  • Login Screen returns to login screen

    - by AbeFM
    After many many reboots in a couple days while experimenting with BIOS settings effecting the speed Hardbrake runs at, today I find after a reboot that I have to type in password to log in - ordinarily I have this disabled. When I DO enter my password, it goes to a black screen for a bit, then returns. I can log in as guest, which does the same thing (minus the password) and if I use the wrong password, it complains instead of doing the same. Using the install disc, I see three partitions on my drive, a ~200 MB boot sector, and two 32 GB (one extended) which seem to share the rest of the SSD. Running FSCK seems to generate tons of errors. The odd bits: All my background stuff is running - I can access stuff served by Subsonic, and see network shares from my windows machines. I can log in in another terminal and do stuff... I just can't get into the GUI/OS proper. Sort of at a loss where to start. Would be happy to free drive of errors if I could (I've another machine, I could mount drive over USB and check it), but it seems everything else is working? edit: Screensaver also seems to kick on, even from fsck's run from the boot menu. i3-2100t, H67 chipset I believe, 12.10, everything's been working fine for the better part of a year. Seen several similar topics, but either they turn out to be something unrelated (fresh install or known graphics issues) or there are no answers. I'm happy to get any logs/info anyone want.

    Read the article

  • Roundcube access from outside local network

    - by Mike K
    I have setup hmailserver on a windows xp pro computer and everything is working great when it comes to getting it to send and receive email. Now here is where my problem comes in... I have setup roundcube using uniserver and everything works great internally. Now, what I am trying to do is to be able to access roundcube from the internet outside of the local network. I have opened up ports 80, 443, and the mysql port yet I am still unable to access this from outside of the network. The error message that I am getting is as followed: Forbidden You don't have permission to access /webmail/ on this server. Additionally, a 404 Not Found error was encountered while trying to use an ErrorDocument to handle the request. I believe something needs to be edited in the apache config file. Any help would be greatly appreciated. I am hoping I dont need to run this on windows server but I'm kinda losing hope and thinking I will need to because I need IIS.

    Read the article

  • Reasons why mod_jk wouldn't work and how to trace them

    - by Bozho
    I've been using one server, then I reinstalled everything on another server, and the mod_jk stopped working. Here is the situation: apache 2.0 sitting "in front" mod_jk used to connect to the apache to tomcat tomcat 6.0.26 used to server the actual requests I followed this tutorial. The result is: accessing http://mysite.com opens the index.html in /var/www/ accessing http://mysite.com:8080/ works OK the logs at /var/logs/apache2 show everything is OK: [Mon Mar 29 22:01:53.310 2010] [28349:3075389184] [info] init_jk::mod_jk.c (2830): mod_jk/1.2.26 initialized [Mon Mar 29 22:01:53 2010] [warn] No JkShmFile defined in httpd.conf. Using default /var/log/apache2/jk-runtime-status [Mon Mar 29 22:01:53 2010] [notice] Apache/2.2.9 (Debian) mod_jk/1.2.26 configured -- resuming normal operations I compared the server.xml, jk.conf, sites-enabled/mysite from the new server to those from the old one and they are identical. The domain name is the same (I updated the DNS record today, and it has refreshed successfully) So the question is, what can go wrong? Is there another place where problems would be logged, if such occur?

    Read the article

  • What's the benefit of object-oriented programming over procedural programming?

    - by niko
    I'm trying to understand the difference between procedural languages like C and object-oriented languages like C++. I've never used C++, but I've been discussing with my friends on how to differentiate the two. I've been told C++ has object-oriented concepts as well as public and private modes for definition of variables: things C does not have. I've never had to use these for while developing programs in Visual Basic.NET: what are the benefits of these? I've also been told that if a variable is public, it can be accessed anywhere, but it's not clear how that's different from a global variable in a language like C. It's also not clear how a private variable differs from a local variable. Another thing I've heard is that, for security reasons, if a function needs to be accessed it should be inherited first. The use-case is that an administrator should only have as much rights as they need and not everything, but it seems a conditional would work as well: if ( login == "admin") { // invoke the function } Why is this not ideal? Given that there seems to be a procedural way to do everything object-oriented, why should I care about object-oriented programming?

    Read the article

  • Cable Internet connectivity Problem?

    - by LightHeaded
    I just got Internet cable from Time Warner/ EarthLink. For some reason I tried everything that Time Warner and Earthlink techs told me to do but nothing seems to work. I still can't connect to the Internet. For some reason they think it's the IPv4 address since it begins with 169. I did everything the told me to 8 times, but they both give me the run around tell me it's the other's fault. How can I fix this once and for all? I have no router. I use cable to connect to the Internet. Windows Vista Cable broadband.

    Read the article

  • WHEA_UNCORRECTABLE_ERROR BSOD - what is wrong?

    - by Arcane
    Hello I have a problem with my PC getting the WHEA_UNCORRECTABLE_ERROR bsod. It is a few months old pc. I am not overclocking or changing anything in bios, but sometimes (this is my third time) I get this bsod out of nowhere. This last time I was just watching some stream on twitch when the pc suddenly crashed. I have everything up to date in windows update and every other driver as well. Everything in Open hardware Monitor seems ok. Right now as of 10 min after the crash, the CPU temp is at 38°C, fans are at 1400RPM I am using Windows 8.1 Pro 64b. Here are my spec: Intel i5-4570 CPU @ 3.20GHz, Kingston HyperX 8GB DDR3 1600 XMP, GeForce GTX 660 with 340.52 drivers, Corsair VS450 power supply, Gigabyte GA-H87-HD3 you can see the BSOD dump here: https://www.dropbox.com/s/88i1q2nnqnpzux1/082514-29656-01.dmp?dl=0 Can anybody tell me what is wrong?

    Read the article

  • Mysql not loading correctly

    - by mcondiff
    PHP 5.3.2 Apache 2.2.15 Mysql 5.1.X Windows XP SP3 I have now configured everything correctly but get a timeout when trying to connect to Mysql via PHP. So frustrated. I don't get an error message, the script just times out. I have made sure I have the correct paths. Fatal error: Maximum execution time of 60 seconds exceeded Any idea why this might be happening? I do a php -v from the command line and everything is normal, no errors. i upgraded PHP from 5.2.6 to 5.3.2 - does there seem to be problems or bugs with this? I am essentially using my previous PHP.ini while editing paths. I am lost. Help! If you need anything from phpinfo() or httpd.conf or php.ini let me know. else

    Read the article

  • What are the disadvantages of automated testing?

    - by jkohlhepp
    There are a number of questions on this site that give plenty of information about the benefits that can be gained from automated testing. But I didn't see anything that represented the other side of the coin: what are the disadvantages? Everything in life is a tradeoff and there are no silver bullets, so surely there must be some valid reasons not to do automated testing. What are they? Here's a few that I've come up with: Requires more initial developer time for a given feature Requires a higher skill level of team members Increase tooling needs (test runners, frameworks, etc.) Complex analysis required when a failed test in encountered - is this test obsolete due to my change or is it telling me I made a mistake? Edit I should say that I am a huge proponent of automated testing, and I'm not looking to be convinced to do it. I'm looking to understand what the disadvantages are so when I go to my company to make a case for it I don't look like I'm throwing around the next imaginary silver bullet. Also, I'm explicity not looking for someone to dispute my examples above. I am taking as true that there must be some disadvantages (everything has trade-offs) and I want to understand what those are.

    Read the article

  • dpkg E: Sub-process /usr/bin/dpkg returned an error

    - by user81269
    I decided to shift around my partitions on my hard drive for a fresh install of Kubuntu. I booted my Ubuntu 10.10 live disc, shifted everything around and attempted to install grub and it didn't work, so I burnt an Ubuntu 12.04 disc and installed it. I got the computer working and wanted to install some packages, but didn't have an internet connection at the time. So (I know this was stupid) I got some debs from previous versions of Ubuntu, as I needed my music, and the other install took a long of time to boot. Once I got my internet connection back, everything worked ok, for a little while. Then I stumbled upon this problem after removing ten broken packages using synaptic: drhax@Spamotard:~$ sudo apt-get install -f Reading package lists... Done Building dependency tree Reading state information... Done The following packages will be REMOVED: libgtk2.0-cil 0 upgraded, 0 newly installed, 1 to remove and 417 not upgraded. 1 not fully installed or removed. After this operation, 2,638 kB disk space will be freed. Do you want to continue [Y/n]? y (Reading database ... 103052 files and directories currently installed.) Removing libgtk2.0-cil ... E: File does not exist: /usr/share/cli-common/packages.d/policy.2.6.gtk-dotnet.installcligac dpkg: error processing libgtk2.0-cil (--remove): subprocess installed post-removal script returned error exit status 1 Errors were encountered while processing: libgtk2.0-cil E: Sub-process /usr/bin/dpkg returned an error code (1) Help would be appreciated. This is my first post, but I do know fair bit about Ubuntu, so feel free to point out any stupid mistakes I have made.

    Read the article

  • Recursive reset file permissions on Windows

    - by Peter Horvath
    There is a big, complex directory structure on a relative big NTFS partition. Somebody managed to put very bad security privileges onto it - there are directories with randomly given/denied permissions, etc. I already run into permission bugs multiple times, and I found insecure permission settings multiple times (for example, write permissions for "Everyone", or false owners). I don't have time to check everything by hand (it is big). But luckily, my wishes are very simple. The most common: read/write/execute on anything for me, and maybe read for Everyone. Is it possible to somehow remove all security data from a directory and giving my (simple) wishes to overwrite everything there? On Unix, I used a chown -R ..., chmod -R ... command sequence. What is its equivalent on Windows?

    Read the article

< Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >