Search Results

Search found 5313 results on 213 pages for 'steve care'.

Page 116/213 | < Previous Page | 112 113 114 115 116 117 118 119 120 121 122 123  | Next Page >

  • can't delete a folder that ends with dot. windows 7 [closed]

    - by Dima
    Possible Duplicate: Can’t delete folder in Windows 7 I got 2 folders that their names end with a dot. both are folders of files I've downloaded through torrent. I can't delete those folders neither access the files inside the folders. Normally I wouldn't care but it is kind of annoying while those 2 folders takes more then 20GB... few people told me that maybe some kind of cmd commands can help with it but I don't really know how to delete the folders through cmd. the location of the folders is E:\tixati\downloads\folder thanks for anyone that would help

    Read the article

  • Do hard drives have to be mounted somewhere?

    - by TheLQ
    I'm creating a cheap JBOD box with ~7 IDE hard drives. However when you take into account 3 existing drives + 2 CD drives that are installed as well, I am out of mounting places inside the server. Will I adversely affect or harm the drives if I just stack them on top of each other in the box? Before you ask if they are going to fall over, this box won't be moved and won't really be touched while its in use. Heat (should) be taken care of by the fans. However I am worried about the vibrations that each drive will put out. Is this an actual issue?

    Read the article

  • how to compare files/directories of 2 separate solaris boxes ?

    - by chz
    Hi Friends I have 2 solaris boxes and I need to check certain directories (on local filesystem and mounted nfs) to make sure that they match up on both boxes and to delete or move the other mismatches to elsewhere on the local filesystem. I investigated for unix commands like rsync, and tree but it appears that these commands are not supported on my Solaris boxes. What is the best approach to this problem with the least pain to solve it ? to use rsync, tree and then diff the outputs or find ? I have trouble limiting the find command to certain directories as there are mounted folders that contain too many xml files that I don't care to much in that directory. What's the find command to search multiple directory paths on a single find command. Thanks Sincerely

    Read the article

  • Install Mozilla Thunderbird on NAS [on hold]

    - by user2295350
    I have a small office with 3 computers running Linux Ubuntu. All the data are stored in a NAS drive (WD MyBook Live). I would like to - somehow - be able to access the email archives as well from each of these computers. For now, I don't really care if it would be possible for each user to access the email archives simultaneously (although that would be ideal). I have read about Portable Thunderbird over Wine as well as installing IMAP on the NAS drive. However, not only I haven't found a complete guide on how to do that but I am also not sure whether either of these solutions would work for me anyway.

    Read the article

  • Haskell: monadic takeWhile?

    - by Mark Rushakoff
    I have some functions written in C that I call from Haskell. These functions return IO (CInt). Sometimes I want to run all of the functions regardless of what any of them return, and this is easy. For sake of example code, this is the general idea of what's happening currently: Prelude> let f x = print x >> return x Prelude> mapM_ f [0..5] 0 1 2 3 4 5 Prelude> I get my desired side effects, and I don't care about the results. But now I need to stop execution immediately after the first item that doesn't return my desired result. Let's say a return value of 4 or higher requires execution to stop - then what I want to do is this: Prelude> takeWhile (<4) $ mapM f [0..5] Which gives me this error: <interactive:1:22: Couldn't match expected type `[b]' against inferred type `IO a' In the first argument of `mapM', namely `f' In the second argument of `($)', namely `mapM f ([0 .. 5])' In the expression: takeWhile (< 4) $ mapM f ([0 .. 5]) And that makes sense to me - the result is still contained in the IO monad, and I can't just compare two values contained in the IO monad. I know this is precisely the purpose of monads -- chaining results together and discarding operations when a certain condition is met -- but is there an easy way to "wrap up" the IO monad in this case to stop executing the chain upon a condition of my choosing, without writing an instance of MonadPlus? Can I just "unlift" the values from f, for the purposes of the takeWhile? Is this a solution where functors fit? Functors haven't "clicked" with me yet, but I sort of have the impression that this might be a good situation to use them. Update: @sth has the closest answer to what I want - in fact, that's almost exactly what I was going for, but I'd still like to see whether there is a standard solution that isn't explicitly recursive -- this is Haskell, after all! Looking back on how I worded my question, now I can see that I wasn't clear enough about my desired behavior. The f function I used above for an example was merely an example. The real functions are written in C and used exclusively for their side effects. I can't use @Tom's suggestion of mapM_ f (takeWhile (&lt;4) [0..5]) because I have no idea whether any input will really result in success or failure until executed. I don't actually care about the returned list, either -- I just want to call the C functions until either the list is exhausted or the first C function returns a failure code. In C-style pseudocode, my behavior would be: do { result = function_with_side_effects(input_list[index++]); } while (result == success && index < max_index); So again, @sth's answer performs the exact behavior that I want, except that the results may (should?) be discarded. A dropWhileM_ function would be equivalent for my purposes. Why isn't there a function like that or takeWhileM_ in Control.Monad? I see that there was a similar discussion on a mailing list, but it appears that nothing has come of that.

    Read the article

  • How to restrict deletion of a folder on NTFS share, but still allow modify access within folder

    - by thinkdreams
    I am setting up a set of scan folders from a scanning copier device, and would like to know the best way to protect the folders (for each department) from moving or deletion, but yet still allow access for the users to modify (i.e. create/add/delete) the scanned files within the folder. Structure is: Share Name Departmental Folder User files The writing of the files initially is taken care of by a service account which has full control. We'd just like to ensure the users cannot accidentally delete the folder (which has already happened) containing all the files, etc. This is for a Windows 2003 server, NTFS permissions. Suggestions would be most appreciated.

    Read the article

  • iPhone VPN to iTunes Home Sharing [migrated]

    - by Philip Crumpton
    my goal is to set up a VPN on a Windows machine that contains my iTunes library then connect to that VPN with my iPhone and be able to utilize Home Sharing remotely. I have read that this can easily be set up if the iTunes library is on a Mac (Network Beacon and YazSoft ShareTool are two products I quickly found). I can't find anyone who has had success on a Windows machine, though. In my thinking, there are two options (aside from buying a Mac): 1.) Existing utility that takes care of this for me (like the Mac-only options listed above) and is compatible with iPhone (Hamachi is NOT compatible with iPhone VPN) 2.) Manually configure a VPN to allow Bonjour multicast (can't find any information on this) FYI my router is a Linksys WRT54GL running Tomato 1.28

    Read the article

  • Is a Windows 7 professional upgrade over vista home possible ?

    - by Saulius
    This has probably been asked before, but I want to be 100% sure before buying: if I have a laptop with an OEM version of Windows Vista Home Premium installed, can I buy the Windows 7 professional upgrade and install it ? I don't care whether it will be an actual upgrade or a clean install (that would actually be even better), what I'm curious is whether this is actually legal (going from home to professional). And if this is possible, how does it work ? Do I have to have Vista installed when upgrading so that the installer can check that I actually do own a genuine copy or does it ask for the vista key (which I'm not sure I have, since it's an OEM version) ?

    Read the article

  • Why is SQLite3 using covering indices instead of the indices I created?

    - by Geoff
    I have an extremely large database (contacts has ~3 billion entries, people has ~280 million entries, and the other tables have a negligible number of entries). Most other queries I've run are really fast. However, I've encountered a more complicated query that's really slow. I'm wondering if there's any way to speed this up. First of all, here is my schema: CREATE TABLE activities (id INTEGER PRIMARY KEY, name TEXT NOT NULL); CREATE TABLE contacts ( id INTEGER PRIMARY KEY, person1_id INTEGER NOT NULL, person2_id INTEGER NOT NULL, duration REAL NOT NULL, -- hours activity_id INTEGER NOT NULL -- FOREIGN_KEY(person1_id) REFERENCES people(id), -- FOREIGN_KEY(person2_id) REFERENCES people(id) ); CREATE TABLE people ( id INTEGER PRIMARY KEY, state_id INTEGER NOT NULL, county_id INTEGER NOT NULL, age INTEGER NOT NULL, gender TEXT NOT NULL, -- M or F income INTEGER NOT NULL -- FOREIGN_KEY(state_id) REFERENCES states(id) ); CREATE TABLE states ( id INTEGER PRIMARY KEY, name TEXT NOT NULL, abbreviation TEXT NOT NULL ); CREATE INDEX activities_name_index on activities(name); CREATE INDEX contacts_activity_id_index on contacts(activity_id); CREATE INDEX contacts_duration_index on contacts(duration); CREATE INDEX contacts_person1_id_index on contacts(person1_id); CREATE INDEX contacts_person2_id_index on contacts(person2_id); CREATE INDEX people_age_index on people(age); CREATE INDEX people_county_id_index on people(county_id); CREATE INDEX people_gender_index on people(gender); CREATE INDEX people_income_index on people(income); CREATE INDEX people_state_id_index on people(state_id); CREATE INDEX states_abbreviation_index on states(abbreviation); CREATE INDEX states_name_index on states(name); Note that I've created an index on every column in the database. I don't care about the size of the database; speed is all I care about. Here's an example of a query that, as expected, runs almost instantly: SELECT count(*) FROM people, states WHERE people.state_id=states.id and states.abbreviation='IA'; Here's the troublesome query: SELECT * FROM contacts WHERE rowid IN (SELECT contacts.rowid FROM contacts, people, states WHERE contacts.person1_id=people.id AND people.state_id=states.id AND states.name='Kansas' INTERSECT SELECT contacts.rowid FROM contacts, people, states WHERE contacts.person2_id=people.id AND people.state_id=states.id AND states.name='Missouri'); Now, what I think would happen is that each subquery would use each relevant index I've created to speed this up. However, when I show the query plan, I see this: sqlite> EXPLAIN QUERY PLAN SELECT * FROM contacts WHERE rowid IN (SELECT contacts.rowid FROM contacts, people, states WHERE contacts.person1_id=people.id AND people.state_id=states.id AND states.name='Kansas' INTERSECT SELECT contacts.rowid FROM contacts, people, states WHERE contacts.person2_id=people.id AND people.state_id=states.id AND states.name='Missouri'); 0|0|0|SEARCH TABLE contacts USING INTEGER PRIMARY KEY (rowid=?) (~25 rows) 0|0|0|EXECUTE LIST SUBQUERY 1 2|0|2|SEARCH TABLE states USING COVERING INDEX states_name_index (name=?) (~1 rows) 2|1|1|SEARCH TABLE people USING COVERING INDEX people_state_id_index (state_id=?) (~5569556 rows) 2|2|0|SEARCH TABLE contacts USING COVERING INDEX contacts_person1_id_index (person1_id=?) (~12 rows) 3|0|2|SEARCH TABLE states USING COVERING INDEX states_name_index (name=?) (~1 rows) 3|1|1|SEARCH TABLE people USING COVERING INDEX people_state_id_index (state_id=?) (~5569556 rows) 3|2|0|SEARCH TABLE contacts USING COVERING INDEX contacts_person2_id_index (person2_id=?) (~12 rows) 1|0|0|COMPOUND SUBQUERIES 2 AND 3 USING TEMP B-TREE (INTERSECT) In fact, if I show the query plan for the first query I posted, I get this: sqlite> EXPLAIN QUERY PLAN SELECT count(*) FROM people, states WHERE people.state_id=states.id and states.abbreviation='IA'; 0|0|1|SEARCH TABLE states USING COVERING INDEX states_abbreviation_index (abbreviation=?) (~1 rows) 0|1|0|SEARCH TABLE people USING COVERING INDEX people_state_id_index (state_id=?) (~5569556 rows) Why is SQLite using covering indices instead of the indices I created? Shouldn't the search in the people table be able to happen in log(n) time given state_id which in turn is found in log(n) time?

    Read the article

  • Reset my windows server 2003

    - by Tim Thoirp
    I was recently given an HP Proliant Server from a friend as a gift. It has Windows Server 2003 installed on it. When I go to boot the system however to log in to Windows Server 2003 it requires an Admin password. I can't figure out the password and my friend doesn't know it either as it has been years since hes used the machine. I don't care about any of the data on the machine I just want to have a new clean version of Windows server 2003 running on it. Any advice would be helpful? And no I don't want to pay for a password cracking tool. Thanks

    Read the article

  • How to restrict deletion of a folder on NTFS share, but still allow modify access within folder

    - by thinkdreams
    I am setting up a set of scan folders from a scanning copier device, and would like to know the best way to protect the folders (for each department) from moving or deletion, but yet still allow access for the users to modify (i.e. create/add/delete) the scanned files within the folder. Structure is: Share Name Departmental Folder User files The writing of the files initially is taken care of by a service account which has full control. We'd just like to ensure the users cannot accidentally delete the folder (which has already happened) containing all the files, etc. This is for a Windows 2003 server, NTFS permissions. Suggestions would be most appreciated.

    Read the article

  • Why are the prices for broadband bandwidth at data centers much higher than consumer/small business offerings?

    - by odemarken
    The prices for broadband bandwidth at data centers are sometimes as much as 10x higher than for a typical small business/consumer connection, at least where I live. Now, I understand those are two differend kind of products, but what exactly are the differences? Is it mainly because the bandwidth you get at a data center is guaranteed (CIR), while a consumer offer lists maximal bandwidth (EIR/MIR)? Or are there other factors as well? (Note: my previous, much more specific question on the same general topic was closed as not constructive. I tried to extract the core issue and present it in a way that can be answered objectively. If you feel that this question is still bad and should be closed, please care to comment and explain why.)

    Read the article

  • Is putting swapfile & temp folder in ramdisk a good idea in Windows 7 64 bit with lots of ram?

    - by Tony_Henrich
    I want my Windows to run as fast as possible. If I have 12GB ram in Windows 7 64bit, quad core cpu, and all apps fit in memory, will the swap file ever be used for anything? The question is about if its a good idea to put the swap file in a ram disk. Would a ram disk help in any way or will Windows intelligently use all the available memory for all its work? I am also thinking of putting the temp folder in a ram disk. I know ram disk is volatile memory and I don't care about its content in it if it gets lost.

    Read the article

  • Power outage during disk wipe. What do I do now?

    - by Mark Trexler
    I was using Roadkil Diskwipe on an external hard drive and the power went out. I had removed it from any outlet connection by the time power was restored to prevent power-spike damage (it's on a surge protector, but I didn't want to rely on that). My question is, where do I go from here? Obviously I don't care about preserving any data currently on it, I just want to make sure the drive itself is not terminally damaged. I'm running chkdsk (full), but I don't know if that's the correct step to assessing any damage. If it makes any difference, the hard drive was unallocated at the time of the outage, as Diskwipe requires that for it to run. Also, could something like this cause latent problems with the drive itself (i.e. serious issues that I won't be aware of when testing it now). I'd appreciate any program recommendations if chkdsk is not the most appropriate diagnostic route. Thank you.

    Read the article

  • Is there a tool for verifying the contents of a Zip archive against the source directory's contents?

    - by Basil
    Here's the scenario: I create a ZIP archive using some GUI package like WinZip, 7-Zip or whatever by right-clicking on a directory "somename" and selecting "Compress to archive 'somename.zip'" When the archive is completed, I open it and discover that some files don't exist in the archive (for reasons yet unknown). I want to find all files that are missing from the archive without having to extract the archive to another directory, then doing directory diff, etc. So.. Is there a tool (GUI or command-line, standalone or built into a compressor, for Windows or Linux, I don't care) that can walk through an archive and compare its contents against a directory on the filesystem?

    Read the article

  • Restrict subversion to only allow certain functions

    - by Farseeker
    I'm in a bit of a situation. We have our private subversion server that we use for development, but one of our government clients is requesting access to our commit logs so that they can get an up-to-date picture of what we've been doing on the system. I don't have a problem with them reading our commit logs, but what I do have a problem with is them having access to our source code - they can't have read or write. The obvious solution is to do an svn log ourselves and give them an export, but they want direct SVN access as they apparently have an auditing solution that will import the svn log command automagically. So, is there a way I can set up access to a subversion repo and deny them access to everything except svn log? I don't care if I have to set up a virtualhost just for this, but it has to be done over http(s). We're also using LDAP for authentication if that makes any difference.

    Read the article

  • Installed Ubuntu 11.10, getting hard disk health warnings

    - by Brad
    I'm getting hard disk health problem warnings... When I click the "examine" button the disk utility pops up. None of my drives are reporting any major issues, and the very first drive doesn't even have a SMART button. I don't really care if one of the drives is crashing, I've got everything backed up, but I just want to know how to stop these god forsaken message boxes from popping up randomly. I have already gone into the Disk Utility and checked "do not notify me if this drive is failing" on all of them except the one that doesn't have the SMART button. I've googled about as much as I can for one day.

    Read the article

  • Comparing two specific properties of a CSV using Compare-Object isn't giving the expected results

    - by MDMarra
    I have a list of users from two separate domains. These lists are in CSV format and I only care about the SAMAccountName, which is a field in these CSVs. The code that I'm working with is currently: $domain1 = Import-CSV C:\Scripts\Temp\domain1.xxx.org.csv | Select-Object SAMAccountName $domain2 = Import-CSV C:\Scripts\Temp\domain2.xxx.org.csv | Select-Object SAMAccountName Compare-Object ($domain1) ($domain2) This is returning only a handful of results (which aren't accurate) in this format: @{samaccountname=SomeUser} => Obviously, Compare-Object isn't evaluating the objects as strings. How do I make this work?

    Read the article

  • Help: My SSL Certificate expired but I can't renew this weekend, is there a way I can disable SSL through IIS?

    - by shogun
    Or perhaps force redirect to HTTP when they request HTTPS? I tried removing the SSL port settings under 'Advanced' but it broke the login page. I don't want to do a deploy/recompile right now but I am able to edit the VIEWS. However I am thinking there may be a way for IIS to do this, but then again I think the .NET code tries to force SSL. Crap. This is bull crap because I was bugging people about getting the new certificate since December and they were like oh we will take care of it... I think it may have even gotten to the point where they were getting annoyed with me hassling them about it! (rant over.) And yes, the key is that more support tickets will be caused by the browser giving a security error than if SSL was removed entirely for one day. Becaues they wouldn't very very likely not notice it being gone.

    Read the article

  • Forcing Windows to be 'awake' during certain hours?

    - by Mun
    I'm running Windows 7 and have the power options currently set up to put the machine to sleep after 2 hours of inactivity. Is there any way to force the machine not to sleep between certain hours, or to automatically wake up? I'm assuming this can't be done directly from within the Windows power settings, but is there a third party utility (preferably free) that could take care of this? I've got backups scheduled every morning at 3-4am, and these don't run if the machine is suspended, so I just need to make sure that the everything is powered up during this time.

    Read the article

  • Cannot access internal network on OSX 10.6.6

    - by cabuki
    Last week, I began having trouble connecting to our internal web servers. Usually, a refresh would take care of it or switching to a different wireless network, but as of yesterday, this wasn't enough. We have an internal DNS server using dnsmasq and a private internal host name (us.lcl). Once I started having more issues with the names not resolving, I tried pinging the server. Using the internal host name (s1.us.lcl), it failed. I tried using the IP address, but that also failed. I have no problems accessing external sites with the exception of it being a bit slower than normal. A reboot yesterday at lunch time after following the instructions here seemed to fix the issue, but when I came into the office this morning, it had stopped working. As of this posting, I cannot ping, ssh or access the web server using the internal host name or ip address. I'm the only one running 10.6 in my office and none of my colleagues has this issue.

    Read the article

  • Is my laptop good enough to support my development needs? [closed]

    - by KodeSeeker
    I have an ASUS Pentium-R Dual Core CPU running at 2.20Ghz. It has 4 gb of built in ram, currently running a 64 bit Windows 7 . I just started graduate school and Im wondering whether I should go in for a new laptop or just repair the nagging battery on my current one. My requirements include - -Ability to support IDE's - I may end up running Eclipse, Visual Studio's and the like to help with my work. - Ability to run multiple VM's (not concurrently). Im currently running a Ubuntu 12 and 9 as VM's (not sure if this is overloading the system) - I'm a non gamer so I really dont care about a minor glitch caused by running a uber heavy game. -In addition I will have heavy use of Office Application Software and will be using my computer to watch movies and stream media. Looking forward to your replies and suggestions!

    Read the article

  • non-interactive ssh sudo... prompts for the password in plain text

    - by Iain
    I'm running some non-interactive ssh commands. The ssh authentication is taken care of fine through the ssh agent, but if I run a command that requires sudo then the password prompt in my terminal is plain text. For example: ssh remotemachine "sudo -u www mkdir -p /path/to/new/folder" will prompt me for the password in plain text. Does anyone know how I can get it to use the normal secure prompt or that I can pass the password via a switch? (as then I can set up a secure prompt on this side before I send the command) Any help is much appreciated.

    Read the article

  • Automatic repair software

    - by ADOConnection
    Do anyone knows any kind of apps or services for "taking care of servers"? (besides managed servers) There are hundreds of ways your server or application can stop working properly. Small things are easy to miss but usually easy to fix. Log overgrouth, configuration issues, etc. Of course there are best practice checklists, but its not a human task to check configuration best practices. Im sure it can be automated: some kind of agent can monitor all system settings, say what is right and wrong and give suggestions on how to make it right. I have to admin several servers and I need some kind of overview of overall situation. As well as a tool, that will fix problems automatically. Can you people suggest something? (I know its a little bit out of rules of SF, but I think this particular question is quite specific) It would be great to have something like http://stackoverflow.com/questions/1451319/asp-net-mvc-view-engine-comparison but for automation software.

    Read the article

  • How do I know I'm running inside a linux "screen" or not?

    - by Jun Chen
    The "screen" refers to a program mentioned in How to reconnect to a disconnected ssh session . That is a good facility. But there is a question I'd really like to know. How do I know whether I'm running inside a "screen"? The difference is: If yes, I know I can safely close current terminal window, e.g., close a PuTTY window, without losing my shell(Bash etc) session. If no, I know I have to take care of any pending works before I close the terminal window. Better, I'd like this status to be displayed in PS1 prompt so that I can see it any time automatically.

    Read the article

< Previous Page | 112 113 114 115 116 117 118 119 120 121 122 123  | Next Page >