Search Results

Search found 6992 results on 280 pages for 'exist'.

Page 124/280 | < Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >

  • nmap reports host up when it isn't

    - by martianway
    On an Ubuntu VM I ran: sudo nmap -sP 192.168.0.* This returned: Starting Nmap 5.00 ( http://nmap.org ) at 2010-12-28 22:46 PST Host 192.168.0.0 is up (0.00064s latency). Host 192.168.0.1 is up (0.00078s latency). Host 192.168.0.2 is up (0.00011s latency). . . . Host 192.168.0.254 is up (0.00068s latency). Host 192.168.0.255 is up (0.00066s latency). The problem is I only have 4 live machines on 192.168.0.* so why did nmap report every ip in the subnet has a live host? The ip address of the Ubuntu machine is 192.168.28.131 From this VM I can ping the live systems on my internal subnet 192.168.0.* and get the expected response. And if I ping a machine that doesn't exist I can get no response as expected.

    Read the article

  • Is there an encrypted write-only file system for Linux?

    - by Grumbel
    I am searching for an encrypted file system for Linux that can be mounted in a write-only mode, by that I mean you should be able to write/append files, but not be able to read the files you have written. Access to the files should only be given when the filesystem is mounted via a password. The purpose of this is to write log files and such, without having the log files themselves be accessible. Does such a thing exist on Linux? Or if not, what would be the best alternative to create encrypted log files? My current workaround consists of simply piping the data through gpg --encrypt, which works, but is very cumbersome, as you can't get easy access to the file system as a whole, you have to pipe each file through gpg --decrypt manually.

    Read the article

  • Is it possible to have a scheduled task click a button in Windows?

    - by Jeremy Lattimore
    One of my clients has an application that can not be run from command line or via API but needs to be run on an scheduled basis. It would require that I launch the application if it's not already running and then click a button. This task would exist on a server. I had read about AutoHotKey but it doesn't sound like it can click buttons from a Scheduled Task. Does anyone have ideas on how to do this? NOTE: As I answered below... I know this is a hack. It would be better to have an app that can actually be scheduled to run. Unfortunately (for me) this application isn't going to be replaced as it is connected to the client's beloved inventory management system and the entire deal is proprietary. So I just need to find an option to attempt this (might not be completely reliable I know). Thanks.

    Read the article

  • USB Device With Embedded Fileserver

    - by Richard Martinez
    I'm attempting to access logs from a proprietary hardware box with no reasonable hope of modifying the software. There is a process on the device to dump log files to a flash drive on the USB port after entering a code sequence. Currently, analysis of the logs requires the following: Physical presence at the device Manual entry of the code sequence Removal of USB device Insertion of USB device into a normal Linux box I'm hoping there is some sort of device that can act as a USB mass storage device but simultaneously make it's contents available as a network file share (wired preferred). Does such a device currently exist? A combo hardware/software solution would also work.

    Read the article

  • Migrating domains - 301 Redirect of all contents of directory

    - by Trufa
    I need to do a 301 redirect with apache since I'm migrating domains. What I would need to do is the following, from certain directories, redirect all of it's content to a different damin (where the file already exists). Let's say I have one.com/files/something.doc or one.com/files/other.php I have already "copied" or "backed up" all the contents of the directory, so the following already exist: two.com/old/files/something.doc and two.com/old/files/other.php So I would just need to redirect anything in the directory "files" (or whatever). I hope the question is clear enough, if not please ask for any clarification needed!! Thanks in advance!!

    Read the article

  • Material to use for computer system cover against UV and salty air?

    - by hippietrail
    I live right next to the sea and have a large window quite close to my computer setup which allows a lot of indirect sunlight to enter. I'd like to buy or make a cover for my computer system. From visiting my usual mom & pop computer shop yesterday I got the impression these might not really exist any more. If I make my own I need a material with these qualities: Block or reduce ultraviolet light which can depolymerize plastics (the sun here in Australia is much stronger than in the northern hemisphere). Block salt-laden sea air which can oxidize USB and other connectors. Not cause static electricity when covering or uncovering. Keep dust off of course (-: My setup is a laptop plugged into a wide-screen LCD with a few external drives. So I think I'd want a largish sheet to flop over the whole desk. Are such covers commonly sold these days? What material(s) should I look for which provides the listed attributes?

    Read the article

  • Did bittorrent sync take advantage of multiple peers

    - by NachoChip
    I am currently testing BT sync, and really like it. However, I cannot find much of the technical detail regarding to how it works. I wonder did anyone know that is the program take advantage of multiple peers connection? For example, if I have a File A(FA), which exist in Computer 1 (C1) and Computer 2 (C2). If I set up a sync from C1 and C2 to a folder in Computer 3 (C3), will I get a better sync up time because I can get FA from 2 sources at the same time (of course, assume I have enough downstream)?

    Read the article

  • How should I organize my backups ?

    - by Patrick
    I'm using for the first time rsync to create daily backups of my websites and I was wondering if I should overwrite the previous copy or should I create multiple copies and overwrite only the oldest one ? (I might not have enough space for that, though). I actually have also this question. Let's suppose most of files are accidentally erases.. does rsync delete all these files from the backup space because they don't exist anymore ? How does exactly work in this case ? thanks

    Read the article

  • ISA bus on newer computers

    - by Kevin Ivarsen
    Are there companies that sell new computers that support old ISA bus expansion cards? We have an aging computer running DOS that operates some machinery via an ISA interface board. Updated versions of this board (e.g. PCI, USB) are not available, and I am concerned about the long-term reliability of the 8+ year old computers we currently keep around as backups. If these newer ISA-capable machines exist, are there any general gotchas to be aware of in terms of compatibility with older expansion boards, ability to run DOS, etc.?

    Read the article

  • Determine which version of linux/unix/darwin I have

    - by John
    I have root ssh/terminal access to a linux server. How do I determine which version of centos I have? Some people suggested I run the command cat /etc/redhat-release but I got an error saying file not found. In fact, i'm not entirely sure i'm even using CentOS. That's what some suggested it might be. Here's a list of commands I tried that gave me no file or directory error: cat /etc/*release* cat /etc/*version* cat /proc/*version* cat /proc/*release* Here's a list of linux commands that do not exist: lsb_release: command not found wget: command not found yum: command not found

    Read the article

  • Selenium RC htmlsuite error with IE

    - by Alina
    I am trying to use Selenium RC but i keep getting this error whenever i start the server. "HTML suite exception seen: java.lang.RuntimeException: sessionId 7643 doesn't exist; perhaps this session w as already stopped?" The command I use is java -jar C:\selenium-remote-control-1.0.3\selenium-server-1.0.3\selenium-server.jar -multiwindow -htmlSuite "*iexplore" "https://user1.apps.com/" "C:\TEMP\Selenium Tests\TestSuite1.html" "C:\TEMP\Selenium Tests\results.html" However with the same command if I change *iexplore to *firefox then it works. I need to run the test with IE, please help!Many thanks!!

    Read the article

  • Why is this MySQL FULLTEXT query returning 0 rows when matching rows are present?

    - by Don MacAskill
    I have a MySQL table with 200M rows which has a FULLTEXT index on two columns (Title,Body). When I do a simple FULLTEXT query in the default NATURAL LANGUAGE mode for some popular results (they'd return 2M+ rows), I'm getting zero rows back: SELECT COUNT(*) FROM itemsearch WHERE MATCH (Title, Body) AGAINST ('fubar'); But when I do a FULLTEXT query in BOOLEAN mode, I can see the rows in question do exist (I get 2M+ back, depending): SELECT COUNT(*) FROM itemsearch WHERE MATCH (Title, Body) AGAINST ('+fubar' IN BOOLEAN MODE); I have some queries which return ~500K rows which are working fine in either mode, so if it's result size related, it seems to crop up somewhere between 500K and a little north of 2M. I've tried playing with the various buffer size variables, to no avail. It's clearly not the 50% threshold, since we're not getting 100M rows back for any result. Any ideas?

    Read the article

  • IIS and PHP restrict IO permissions

    - by ULTRA_POROV
    I have php installed trough a fastCGI module. Is there a way to restrict the module (php.exe) read / write permissions to only the directory (+ subdirs) of the IIS site that is calling it? I need this to prevent one IIS PHP site from having access to files outside its own directory. How to do this? Is there a setting in php.ini or in the IIS configuration? I believe such a feature could exist, because when a file on the server is requested the root path of the site is also known, all it would take is that IIS passes this path to the php module, and the php module should on its end allow only IO operations within this path. PS: I know it is possible to achieve this by using a different windows account for each website, this is not an option.

    Read the article

  • Exchange 2007 - Distribution group not work correctly

    - by AlexB
    I have simple server Windows 2008 Standart SP2 with Exchange 2007 with apply SP1. Only one DC and one Exchange server (not exist Edge) in the same server. I have first distribution group by name "[email protected]" with 10 users. Working with internal and external mail correctly. I have second distribution group by name "[email protected]" with 3 users. Working with internal mail correctly, but external mail delivery failed 2 from 3 users. Only 1 users proclaims that external mail delivered correctly. Those 3 users they are too member of distribution group [email protected]" . I am not found in the eventlog any problem with Exchange server. Any user and any distribution group not check the box "Require that all senders are authenticated." Any ideas on how to solve the problem?

    Read the article

  • options for deploying application

    - by terence
    I've created a simple web application, a self-contained tool with a user system. I host it publicly for everyone to use, but I've gotten some requests to allow companies to host the entire application privately on their internal systems. I have no idea what I'm doing - I have no experience with deployment or server stuff. I'm just some person who learned enough JS and PHP to make a tool for my own needs. The application runs with Apache, MySQL, and PHP. What's the best way to package my application to let others run it privately? I'm assuming there's better options than just sending them all the source code. I'd like to find a solution that is: Does not require support to set up (I'm just a single developer without much free time) Easy to configure Easy to update Does there exist some one-size-fits all thing that I can give to someone, they can install it, and bam, now when they go to http://myapplication/ on their intranet, it works? Thanks for your help.

    Read the article

  • Different versions of 2.5" drives? Unscreawing Intel 320 for slimmer drive to fit newer thinkpads?

    - by hhh
    My x60s laptops come with about 1-2mm heigher 2.5" HDD than newer x220 models. This is totally stupid when one would like to reuse the old drive in the newer laptops. I bought newer 2.5" 320 Intel SSDs and they seem to have such 1-2mm gap to unscrew but I am unsure whether it is meant for opening. Could someone instruct what to do here? Look the manufacturer has started changing the old good 2.5" drives into slightly different versions, now it means slow compability issues to fix or totally new 2.5" drives. Ideas how to proceed? Unscrewing newer 320 drives or are some other versions of 2.5" drives meant for x220? Does there exist some sort of racks to get drives working between different comps? Perhaps booting from laplink is currently the easiest solution to get things working when changing harddrives between comps? Perhaps related http://www.zdnet.com/blog/hardware/seagate-announced-super-slim-25in-momentus-thin-hard-drive/6439

    Read the article

  • Restoring openldap configuration from ldif file?

    - by Solignis
    I am trying to restore my OpenLDAP server from an ldif file I made before it went horribly wrong. All of the websites I found talk about using slapadd but when I run slapadd -v -l ~ns01/openldap_config.ldif it returns the error Could not stat config file "/etc/ldap/slapd.conf" : No such file or directory (2) slapadd: bad configuration file! When I look in the directory indeed the file does not exist, but this is because Ubuntu 11.10 is running in RTC mode, my LDAP server uses the /etc/slapd.d/cn=config for the configuration. So? What am I missing? Is slapadd the wrong tool to use? As I said above I running on Ubuntu 11.10 server edition 64-bit.

    Read the article

  • Distributed website server redundancy

    - by Keith Lion
    Assume a website infrastructure is very complicated and is fully distributed (probably like most large web companies). Am I right in thinking that although there are all these extra web servers to handle multiple client requests, there is still a single "machine" whereby users must enter? I am guessing this machine will be the one physically associated to the IP address? I ask because I need to know whether, in places where distributed systems exist, there is still a single point of failure- usually the control node or, in this example, the machine connected to the public internet? Surely there cannot be two machines connected to the internet, as they would have to have different IP addresses? This "machine" may not be a server per se, but maybe it is a piece of cisco equipment. I just need to know whether, in the real world, these distributed systems still have a particular section where they depend on the integrity of one electronic device?

    Read the article

  • Visual Studio 2010 tool to order members inside #region

    - by Michael Swan
    There exist a tool for Visual studio which order alphabetically the members grouped by #region ? means #region A1() ... A2() ... A0() ... B1() ... Z1() ... B5() ... #endregion #region C1() ... C2() ... C0() ... D1() ... X1() ... Y5() ... #endregion With that tool, I want like: #region A0() ... A1() ... A2() ... B1() ... B5() ... Z1() ... #endregion #region C0() ... C1() ... C2() ... D1() ... X1() ... Y5() ... #endregion Thank you

    Read the article

  • Software for defining rules for folder permissions and monitoring deviations

    - by Kjensen
    Let's say a company has a large number of users, and each user has a home area. On each share used for home area folders, I would like to define some rules saying who is supposed to have which permissions on the folder. Then I would like to audit automatically, that this is actually the case and get some sort of report on deviations. So a rule for \MegaServer\Home01 could be defined something like: Domain Admins - Full Control Backup Agent - Read [Home folder owner] - Full Control I am talking about Windows platform and Windows servers, although I think it would most likely also work for *nix machines that expose Windows shares. Does software like this exist? I could roll my own basic version, but if something already exists, that is usually a better option. I am aware of tools to make displaying permissions easier (AccessEnum, DumpSec), but that is not what I am looking for.

    Read the article

  • How to tell if any MySQL connections has been dropped or timed out?

    - by Continuation
    A client is using PHP to connect to MySQL. The PHP scripts and the MySQL database are located on 2 different Linux servers. He complained that database connections were being dropped or timed out and asked me to take a look. Is there any place in MySQL that can show me what and how many connections have been dropped or timed out? I looked into slow query log and didn't see anything. Any suggestions on how to diagnose this dropped/timed out database connection problem? Thanks EDIT: Slow query log is enabled in my.cnf: log-slow-queries=/var/log/mysql-slow-queries.log And when I do a mysql> show global status; I got: | Slow_queries | 11402347 | So there are a lot of slow queries. But the file /var/log/mysql-slow-queries.log doesn't exist. Why is that?

    Read the article

  • Howto dianostic problem in MySQL Cluster?

    - by maj
    Hi, I have set up a MySQL cluster following exactly this howto. Page 1 is completed, but the problem I can see the nodes in ndb_mgm for a little while, and then I get ndb_mgm> show; Cluster Configuration --------------------- [ndbd(NDB)] 2 node(s) id=2 (not connected, accepting connect from 192.168.0.101) id=3 (not connected, accepting connect from 192.168.0.102) [ndb_mgmd(MGM)] 1 node(s) id=1 @192.168.0.103 (Version: 5.1.45) [mysqld(API)] 2 node(s) id=4 (not connected, accepting connect from any host) id=5 (not connected, accepting connect from any host) ndb_mgm> So the questions is; How do I diagnostic this problem? Does there exist log files I can look in?

    Read the article

  • suddenly can't connect to Windows 7 PC from xbmc

    - by Damon
    I have xbmc installed on a softmodded classic xbox and all of a sudden I cannot connect to the shared folders on my Windows 7 PC. Well, it can connect but it wants a username and password, but none exist on my computer. It's been several weeks since I last used it, but it was working fine then! and no particular changes I can think of aside from I guess a few Windows updates. I've seen reference online to Windows Live essentials causing issues with file sharing, but I don't even have that installed. And I made sure to tick 'turn off password protected sharing' in my advanced network setup. I had the guest account turned off and working fine before, and tried turning it on but no luck. I can access my shared folders fine from other Computers and a ps3. I've read a couple threads of similar situations on various folders and none of their solutions have worked.

    Read the article

  • Sending files through ssh

    - by Frion3L
    I need to send files to a server using ssh. I have never used ssh so this is being really frustrating to me. Mention the client (me) is using windows and the server is using Ubuntu. I connected to the server using ssh2 ip, and then loging with an account I have. Now, I would like to send my files to a folder in the server, so, I moved to the folder and I used this command: scp test.txt user_name@host_direction server_folder_destination And it always return that it can't do 'stat' over test.txt, the file doesn't exist, and so. I'm assuming ssh2 can't see the file in my computer root (C:), so I tried to specifie more and added: C:\test.txt, but apear the same error. I don't know what is happening. Any hints please? Thanks

    Read the article

  • How to remove an entry from Chrome's Remembered URLs from the url bar?

    - by cmcculloh
    I've got a url in Chrome "local.mysite.com" that autopopulates when I start typing "local.my" into the URL bar. Note that this URL DOES NOT EXIST in my browser history (at chrome://history/#e=1&p=0) because it isn't a real site and therefor couldn't ever be successfully visited and therefor never shows up in my history. The URL I want is "local.mysite.com/subdir/". That URL is like 3 down in the suggested results because I keep accidentally hitting "enter" when it auto-suggests the unwanted first URL and thus re-enforcing it's assumption that that is the one I want. How do I get rid of the "local.mysite.com" entry in Chrome's memory?

    Read the article

< Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >