Search Results

Search found 56300 results on 2252 pages for 'local working'.

Page 649/2252 | < Previous Page | 645 646 647 648 649 650 651 652 653 654 655 656  | Next Page >

  • Install xenserver on iscsi target

    - by Ghassen Telmoudi
    Is is know that Xen Server is based on CentOS, and it is fact that CentOS can be installed on an ISCSI target flawlessly, but I wanted to install Xen server on an ISCSI target I couldn't find a way. I already tried the latest version of xen server 6.2, and there is no obvious way who to do the installation without a local disk. Does anybody have an idea about the subject, or did someone know how to do it? Please share your experience about this subject.

    Read the article

  • arp "who-has tell" on cloned machine

    - by mcmorry
    I have a urgent problem to solve today, but I'm lost. Please help. I've cloned a Virtual Machine hosted on VM Ware ESXi 4.1 The OS is now Ubuntu Server 12.04 LTS, but at the time of cloning it was 10.04 LTS. I fixed the MAC address manually inside /etc/udev/rules.d/70-persistent-net.rules. It is a known problem on Ubuntu. I had to remove the old MAC address and set the new one as eth0. Everything seems to work fine, except ARP. My provider OVH sent me a warning to resolve it today (this is the second day) or they will block my IP! The log contains many lines like this: Tue Jun 5 01:04:29 2012 : arp who-has 178.32.136.212 tell 178.32.136.224 where .224 is the cloned server that is causing problems, and .212 is the cloned one. arp -na returns: ? (178.33.230.254) at 00:07:b4:00:00:02 [ether] on eth0 ? (178.32.136.212) at 00:50:56:09:8e:f1 [ether] on eth0 The first IP is the ESXi machine. The second one should not be there. I'm not an expert and I don't know what else to do to fix this problem. Any help will be very appreciated. Thanks. EDIT: ifcofig on .224: eth0 Link encap:Ethernet HWaddr 00:50:56:01:32:c6 inet addr:178.32.136.224 Bcast:178.32.136.255 Mask:255.255.255.0 inet6 addr: fe80::250:56ff:fe01:32c6/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:399924 errors:0 dropped:465 overruns:0 frame:0 TX packets:241884 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:58006071 (58.0 MB) TX bytes:663603166 (663.6 MB) lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:516216 errors:0 dropped:0 overruns:0 frame:0 TX packets:516216 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:236284275 (236.2 MB) TX bytes:236284275 (236.2 MB) ifconfig on .212: eth0 Link encap:Ethernet HWaddr 00:50:56:09:8e:f1 inet addr:178.32.136.212 Bcast:178.32.136.255 Mask:255.255.255.0 inet6 addr: fe80::250:56ff:fe09:8ef1/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:16014 errors:0 dropped:0 overruns:0 frame:0 TX packets:14511 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:15134444 (15.1 MB) TX bytes:2683025 (2.6 MB) lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:9944 errors:0 dropped:0 overruns:0 frame:0 TX packets:9944 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:1139347 (1.1 MB) TX bytes:1139347 (1.1 MB)

    Read the article

  • IIS7 80 port outside doesn't work

    - by ihorko
    I have created web site, added to IIS 7, in the binding set up host name as "mysite.com". (here "mysite.com" is my registered domain that points to my IP address) So when I assigned port 8095 and open site as mysite.com:8095 it succesfully opens on my local pc and outsite my network pc, but if I set up port 80 there, http://mysite.com opens only on my pc, but not in outside pc. Firewall is disabled. How to resolve that problem, please help!?

    Read the article

  • postfix error fatal:table lookup

    - by samer na
    here the mail.log server localhost: Can't connect to local MySQL server through socket '/var/run/mysqld /mysqld.sock' (2) Mar 23 23:07:19 ubuntu postfix/trivial-rewrite[6417]: fatal: mysql:/etc/postfix/mysql_virtual_alias_maps.cf(0,lock|fold_fix): table lookup problem Mar 23 23:07:20 ubuntu postfix/smtpd[6401]: warning: problem talking to service rewrite: Success Mar 23 23:07:20 ubuntu postfix/cleanup[6296]: warning: problem talking to service rewrite: Connection reset by peer Mar 23 23:07:20 ubuntu postfix/master[6291]: warning: process /usr/lib/postfix/trivial-rewrite pid 6417 exit status 1 Mar 23 23:07:20 ubuntu postfix/master[6291]: warning: /usr/lib/postfix/trivial-rewrite: bad command startup -- throttling

    Read the article

  • How would I isolate one networked PC to LAN only?

    - by itsraine
    I would like to have one of my PCs available to the rest of my home network for file sharing and VNC access, but I want to block any Internet traffic going to and from the PC. In other words, I want all local PCs I have connected to the router functioning as any normal LAN would, but when it comes to the Internet I want one particular PC to be "safe" from the Internet. My guess is that this is some sort of port blocking or some other router function, but I'm not quite sure.

    Read the article

  • Get WebDav uploading progress and status (Linux, davfs2)

    - by Hnatt
    I am using WebDAV on Linux box with davfs2 1.4.6. When I copy a file to a mounted WebDAV service, it is goes rather fast, just like a regular local drive operation. And it actually is, because the file is first copied to ~/.davf2/cache directory. But how do I know that uploading is finished and where do I see current progress? Is there a way to know that uploading failed due to lack of space or file size restrictions?

    Read the article

  • Windows FTP Client - Think Transmit for Mac

    - by richzilla
    Hi All, Im on the hunt for a decent FTP client for windows. Im looking for something that works in a similar way to transmit for mac. Features im looking for ideally would be the ability to synchronize files between local and server, and most interestingly, to mount an ftp server as a seperate drive. I know i can get individual programs to do each if those, but im looking for an all in one solution... Any ideas?

    Read the article

  • How do you backup 40+ Centos5.5 servers?

    - by John Little
    We are embarrassed to ask this question. Apologies for our lack of UNIX expertise. We have inherited 40+ centos 5.5 servers, and don't know how to back them up. We need low level clone type images so that we could restore the servers from scratch if we had to replace the HDs etc. We have used the "dd" command, but we assume this only works if you want to back up one local disk to another, not 40 servers to one server with an external USB HD attached. All 40 servers have a pair of mirrored disks (dont know if its HW or SW raid). Most only have 100MB used. SErvers are running apache, zend, tomcat, mysql etc. Ideally we dont want to have to shut them down to backup (but could). We assume that standard unix commands like tar, cpio, rsync, scp etc. are of no use as they only copy files, not partitions, all attributes, groups etc. i.e. do not produce a result which can simply be re-imaged to a new HD to get the serer back from dead. We have a large SAN, a spare windows box and spare unix boxes, but these are only visible to one layer in the network. We have an unused Dell DL2000 monster tape unit, but no sw or documentation for it. WE have a copy of symantec backup exec, but we have no budget for unix client licenses. (The company has negative amounts of money). We need to be able to initiate the backup remotely, as we can only access the servers in person in an emergency (i.e. to restore) Googling returns some applications to do this, e.g. clonezilla - looks difficult to install and invasive. Mondo, only seems to support backup if you are local to the machine. Amanda might be an option, but looks like days/weeks of work to learn and setup? Is there anything built into Centos, or do we have to go the route of installing, learning and configuring a set of backup softwares? Any ideas? This must be a pretty standard problem which goggling doesnt give an obvious answer.

    Read the article

  • Sharing files between multiple sites using only desktop software

    - by perlyking
    Our organisation has three sites; a head office, where the master copies of company files are stored, plus two branch offices using only workstations and a NAS or two. Currently we're talking about <10GB. At the main office, we have no admin access to the file server, as this is entirely controlled by the larger institution where we are located. For the same reason, we have no VPN remote access to this network. Instead, we simply have access to a network share using over a Novell LAN. Question: how can we share files between offices in way that minimises latency, i.e. that gives us a mirror of the main network share at each site? (There is little likelihood of concurrent editing, and we can live with the odd file conflict now and again). Up to now branch office staff have had to use GotoMyPC-type solutions to remotely access files held at the main office. Or email. I was hoping to use Google Drive on a dedicated workstation at each office to sync the contents of the network share (head office) or NAS (branch offices) via the cloud, but at my last attempt (29 Jun '12), the Google Drive installer would not allow me to designate the remote network share as the "target" folder. (I chose Google Drive over Drobbox et al. as we already use GMail for corporate mail) The next idea was to use a designated workstation at head office to mirror the network share to a local drive, then use Google Drive to push that to the cloud. This seems a step too far. Nor do I have any good ideas about how to achieve this network/local mirroring, as we can't, for example, install the rsync daemon on the server. I do not want to use Google Drive locally on each workstation as this will inconvenience users, and more importantly, move files off the backed-up, well-maintained (UPS, RAID etc) network share at head office. Our budget is only in the £100's. Should we perhaps just ditch the head office server and use something like JungleDisk? At least this presents the user with what appears to be a mapped drive.

    Read the article

  • IT merger - self-sufficient site with domain controller VS thin clients outpost with access to termi

    - by imagodei
    SITUATION: A larger company acquires a smaller one. IT infrastructure has to be merged. There are no immediate plans to change the current size or role of the smaller company - the offices and production remain. It has a Win 2003 SBS domain server, Win 2000 file server, linux server for SVN and internal Wikipedia, 2 or 3 production machines, LTO backup solution. The servers are approx. 5 years old. Cisco network equippment (switches, wireless, ASA). Mail solution is a hosted Exchange. There are approx. 35 desktops and laptops in the company. IT infrastructure unification: There are 2 IT merging proposals. 1.) Replacing old servers, installing Win Server 2008 domain controller, and setting up either subdomain or domain trust to a larger company. File server and other servers remain local and synchronization should be set up to a centralized location in larger company. Similary with the backup - it remains local and if needed it should be replicated to a centralized location. Licensing is managed by smaller company. 2.) All servers are moved to a centralized location in larger company. As many desktop machines as possible are replaced by thin clients. The actual machines are virtualized and hosted by Terminal server at the same central location. Citrix solutions will be used. Only router and site-2-site VPN connection remain at the smaller company. Backup internet line to insure near 100% availability is needed. Licensing is mainly managed by larger company. Only specialized software for PCs that will not be virtualized is managed by smaller company. I'd like to ask you to discuss both solutions a bit. In your opinion, which is better from the operational point of view? Which is more reliable, cheaper in the long run? Easier to manage from the system administrator's point of view? Easier on the budget and easier to maintain from IT department's point of view? Does anybody have any experience with the second option and how does it perform in production environment? Pros and cons of both? Your input will be of great significance to me. Thank you very much!

    Read the article

  • Windows 7 can't identify network

    - by Carl Hörberg
    I use a Windows 7 machine to share my internet connection, but the one network interface which are connected to my local network is marked as "Unidentified network", the sharing works well anyway but because the interface can't be chosen as Home network i can't use the HomeGroup features etc. Do you know which requirements an interface has meet to identify a network in Windows 7?

    Read the article

  • Recreate/Regenerate vhost config in cpanel for one account

    - by Gabriel
    So my boss as allways was messing with cpanel on our production server (we only have production servers, which is bad enough)... so now all of the accounts is pointing to the default apache page in /usr/local/apache/htdocs instead of /home/useracc/public_html. Is there a way how to tell cpanel/WHM to recreate the vhost configuration of the account from scratch? the account has a database and emails associated with them so before i do anything i don't want to mess anything up with cpanel.

    Read the article

  • Postfix server to receive emails for anyonymous user

    - by sachitad
    I have a postfix server configured with imap. Only the recipient with the user account in the system is accepted. For example: rcpt to: test@localhost will yield the following error: 550 5.1.1 <test@localhost>: Recipient address rejected: User unknown in local recipient table What I want to achieve is setup virtual maps which accepts email to all the users (even if the user doesn't accept in the system) then forward all those emails to a specific user mailbox. Is something like this, possible?

    Read the article

  • Configuring snedmail to forward mail for a specific domain to a specific mail server without using M

    - by aHunter
    I am new to sendmail and would like to configure sendmail to forward all mail for a specific email address to another internal mail server. I need it to ignore the MX records and only send it to the server I specify but am not sure which files to edit or how to configure the sendmail config. Is it sufficiant to add the server to the /etc/hosts and the /etc/mail/local-host-names files? Thanks in advance.

    Read the article

  • Share the same subnet between Internal network and VPN Clients

    - by Pascal
    I would like to set up a configuration where VPN clients connecting to my Forefront TMG can access all the resources of my Internal network without having the to use the option "Use default gateway on remote network" on the VPN's TCP/IP Ipv4 Advanced Settings. This is important to me, since they can use their own internet while accessing my network through VPN (the security implications of this are acceptable on my cenario) My Internal network runs on 10.50.75.x, and I set up Forefront TMG to relay the DHCP of my Internal network to the VPN clients, so they get IPs from the same range as the Internal network. This setup initially works, and the VPN clients use their own internet, and can access anything that is on the internal network. However, after a while, HTTP Proxy Traffic from the Internal network starts getting routed to the IP of the RRAS Dial In Interface, instead of the IP of the Internal's network gateway. When this happens, the HTTP Proxy starts getting denied for obvious reasons. My first question is: does this happen because Forefront TMG wasn't designed to handle a cenario that I described above, and it "loses itself"? My second question is: Is there any way to solve this problem, either through configuration or firewall policies? My third question is: If there's no way that it can work with the cenario above, is there another cenario that will solve my problem, and do what I'd like it to do properly? Below are my network routes: 1 => Local Host Access => Route => Local Host => All Networks 2 => VPN Clients to Internal Network => Route => VPN Clients => Internal 3 => Internet Access => NAT => Internal, Perimeter, VPN Clients => External 4 => Internal to Perimeter => Route => Internal, VPN Clients => Perimeter Tks!

    Read the article

  • postgresql login from remote

    - by Hellnar
    I want to give remote access to my postresql db (8.2) to computers that are at the same lan, at the default config I have added this line to pg_hba.conf where xxx.xx.xx.xx is the ip of the machine that hosts postgresql. This machine is a windows 2k server. # IPv4 local connections: host all all 0.0.0.0/0 password host all all xxx.xx.xx.xx/24 password There no firewall or such blocking the connection between and `listen_addresses = '*' for postgresql.conf .

    Read the article

  • 'Save as' dialogue box in Vista/Win7

    - by DrZaiusApeLord
    Is there a way to get this box to list the local and network drives first instead of listing favorites and libraries first? I have a user who claims this is how this box was until fairly recently but I have not made any changes nor pushed out any big updates. Now that he mentions it, its starting to bug me too. Thanks.

    Read the article

  • Windows: make browsers do a DNS-lookup even when the Computer is offline

    - by leosok
    I use a local DNS-Server (MicroDNS) which I set via netsh to redirect any query to my own page. A little webserver running inside my software answering something like "this page is not whitelisted". It works when connected to the Internet but does not work when offline. The Browsers stop looking up the DNS. How could I make Browsers go to my page, whatever I enter in the address line, WHEN OFFLINE?

    Read the article

  • URIs for QuickTime X

    - by JP
    Apple lets you Stream video over HTTP with their new tech, but is there a URI scheme to open the .m3u8 files from a browser in Quicktime X? (ie. Open as a URL, rather than downloading and attempting to play the local file)

    Read the article

  • Blue screen while using windows even in safe mode

    - by FATE
    I have this problem with Windows 7 after I detected and cleaned up some viruses with Nod32. problem signature: Problem Event Name: BlueScreen OS Version: 6.1.7601.2.1.0.256.48 Locale ID: 1065 Additional information about the problem: BCCode: 50 BCP1: FFFFF900CA578010 BCP2: 0000000000000000 BCP3: FFFFF9600019619B BCP4: 0000000000000002 OS Version: 6_1_7601 Service Pack: 1_0 Product: 256_1 Files that help describe the problem: C:\Windows\Minidump\060312-34195-01.dmp C:\Users\Fatemeh\AppData\Local\Temp\WER-51215-0.sysdata.xml How can I prevent these crashes?

    Read the article

  • How to configure a hostname of sub-application?

    - by BrunoLM
    I have a structure like this: Website |- Controllers |- Models |- Views |- Content |- Static (APP) Website is an application using an ASP.NET 4.0 pool. Static is a sub-application using a not managed application pool. On Website bindings I've set local.domain.com to have access through port 80. I want to access the Static app using static.domain.com, but I don't find the option to configure the binding. How is it possible to configure like that?

    Read the article

< Previous Page | 645 646 647 648 649 650 651 652 653 654 655 656  | Next Page >