Search Results

Search found 21130 results on 846 pages for 'nfs client'.

Page 304/846 | < Previous Page | 300 301 302 303 304 305 306 307 308 309 310 311  | Next Page >

  • While Mail Forwarding with exim, how do I rewrite the To header with true destination address

    - by Jom
    I have mail forwarding setup with exim using a domain forwarding file. virtual_aliases_nostar: driver = redirect allow_defer allow_fail data = ${if exists{/etc/valiases/$domain}{${lookup{$local_part@$domain}lsearch{/etc/valiases/$domain}}}} file_transport = address_file group = mail pipe_transport = virtual_address_pipe retry_use_local_part domains = lsearch;/etc/localdomains unseen It is working fine. However, I would like to rewrite the "to" header. In my system filter, I would like to put something like: headers remove to headers add "To: $recipient:" I've tried: headers remove to headers add "To: $recipient:" headers remove to headers add "To: $h_env-to:" headers remove to headers add "To: $env-to:" The intent is to have the end recipient see their own email address in the To: line of their mail client. I can't seem to figure out what the correct header is for the final destination of the email so that I can put it in the to header. I've read through the Exim docs and can't seem to find it. I've also looked in the headers in an email at a mail client and can't see it there either. Any suggestions would be appreciated.

    Read the article

  • Will search engines reindex a page that has been set to redirect on to a newer site page?

    - by Luke Duddridge
    We were asked by a client to change a website so that any pages/Urls we were hosting on an older site would now redirect to a newer site hosted somewhere else and a different domain name to boot. We did this by changing each page in the IIS site management, to redirect to a url on their new domain instead of rendering a page locally. According to the redirect tool here: http://www.webconfs.com/redirect-check.php . What we have done is search engine friendly. Problem now is... the client has been on a course learning all about meta tags and so thinks they have a better understanding of the "matrix" (remember there is no spoon). As Google still has the older site appearing in a search, this isnt helping matters. I have tried to explain, we have to wait for Google to reindex. I'm not blowing smoke am I? I'm now starting to wonder... will the older site always appear in a search, even though the pages don't exist? Is there a better way I should be redirecting their site to ensure google will stop keeping an index of pages that no longer exist and would instead replace them with the content in the newer site? a suggestion on the site mentioned above is to use the code: Response.Status="301 Moved Permanently" Response.AddHeader "Location","http://www.new-url.com/" Does using the option in the IIS management tool to redirect the url not do the same?

    Read the article

  • Forward emails from specific domain in Exchange

    - by neildeadman
    Our Exchange server handles emails for @ourdomain.com (for example). We have multiple clients that will send emails to our [email protected] email address and we want to configure server-side rules that will forward emails from each client's domain to a different email address within our exchange server. For example: [email protected] sends an email to [email protected] and we forward it to [email protected] [email protected] sends an email to [email protected] and we forward it to [email protected] ...and so on. It would be nice if we can additionally stop the email arriving in the [email protected] mailbox, but that is not a specific requirement. We have a rule setup in Outlook that sort of works, but it doesn't do all from a domain only specific email addresses. It does work when Outlook is not running which is a start. I realise it would be easier to give each client a partiuclar email address and have them email straight to that rather than all use the same, but this is what I have been asked to setup.... :S

    Read the article

  • NIS server setting problem in ubuntu

    - by Asma
    Hi, I set NFS for server-client that works properly. Now I am trying to set NIS server-client on same PC's. I am following the instruction from "SettingUpNISHowTo" from https://help.ubuntu.com/community/SettingUpNISHowTo link. But in step-10, "sudo /etc/init.d/nis restart".......it show error fail. If I try to use "ypcat passwd" to check......it shows error YPBINDPROC_DOMAIN: Domain not bound No such map passwd.byname. Reason: Can't bind to server which serves this domain Can anyone able to help me to get rid out from this problem? Is all the step in the document is proper to configure the NIS server? Thanks in advance.

    Read the article

  • Limit maximum incoming connections to a port using iptables

    - by Harley
    I have a server that has apache listening on a number of ports. Some ports are used for configuring the server, and another is used to download large files. My problem is that when I have a large number of clients downloading files, the web interface is uncontactable. I would like to limit the number of clients connecting on the "large file" port so that apache always has available connections to configure the server. A REJECT is fine, the client trying to download the file will back off and retry later. Each client only has one connection open to the server at a time, so limiting by IP won't work. I know I could put something in front of apache to manage this, but I'd really like to do it in iptables, without adding more software.

    Read the article

  • Baffling Exchange 2010 issue , Outlook 2003

    - by Nick Weber
    Recently we upgraded from Exchange 2003 to Exchange 2010. The Exchange 2003 server is still up and active with a handful of users still there we are planning on migrating. Also we have mobile devices and a Blackberry BES server setup publishing to Exchange 2010 public folders free busy. The unfortunate issue is majority of our clients are still Outlook 2003. The issue seems to be hit and miss but sometime clients cannot view other client’s calendars. They get (cannot refresh free/busy) when they add another user and check their available times. This is also true with booking resources (conference rooms) . One thing we noticed is that the Blackberry BES seems to have created its own free/busy folders, so we have some users who use the BES on free/busy folder and other users who don’t use the BES on other free/busy folders. If any of the clients log-in via “Exchange 2010” web-access or with a Outlook 2007 client they can view the calendaring just fine. Thanks for your help Nick @Toshana

    Read the article

  • Outlook 2010 Error

    - by Trevor Sullivan
    I'm running Outlook 2010 SP1 on Windows 7 x64 SP1, and I'm getting an error message saying "Your Microsoft Exchange administrator has blocked the version of Outlook that you are using. Contact your administrator for assistance." I'm still able to log into my account using Outlook Web Access (OWA), so I know that my account is working just fine. Outlook 2010 with Service Pack 1 is the standard for Windows 7 client systems at this organization, and other people are able to access their e-mail just fine. When my account was initially configured, I was able to use Outlook for a couple of days, and then it suddenly stopped working, providing only the above error message. Do you have any ideas on what I should look into to resolve this problem? Is there any information I can obtain on the client side that will help the Exchange folks investigate the issue further? Is there any verbose logging I can enable, or diagnostic logging in Outlook? Cheers

    Read the article

  • foobar.com working, but www.foobar.com not working?

    - by dpmattingly
    I am setting up a web site for a client. She is using GoDaddy for domain registration, and a hosting company I have never used before. After setting up the nameservers on GoDaddy's side, the address foobar.com (for example) is correctly directing to the new site. However, the address www.foobar.com is redirecting to a 404 page on the hosting company's side. I've been dealing with customer service on the hosting side, and they have told me various things including wait for DNS propagation (which has obviously happened since the 404 page is on their side), and to make sure that the nameservers on GoDaddy's side were entered in lower case instead of upper case (which I know doesn't matter since nameservers are case insensitive). I think I'm getting the runaround from the hosting company, but the client had signed up with them before I came to the project, so if possible I'd like to resolve this issue with them before we start treating it as a loss. Does anybody know what could cause foobar.com to resolve correctly but www.foobar.com to not resolve? How would I best be able to suggest a fix to this through the technical support channels of a hosting company?

    Read the article

  • Does Exchange support plussed users (e.g. [email protected]) or a similar mechanism?

    - by Jens Bannmann
    Sendmail supports a feature called 'plussed users'. Once enabled, emails sent to [email protected], [email protected] and [email protected] are automatically delivered just like mails to [email protected]. There is no need to register or set up these 'plus suffixes'. The user can just use them and set up client-side filtering rules on his own. Does Exchange support a similar mechanism? If so, how to enable it? Note that I don't want answers about other means of filtering, e.g. spam/junk filtering, server-side or client-side rules, email aliases/addresses that are configured explicitly and so on.

    Read the article

  • How can I track all emails sent from my users?

    - by schnapple
    My client runs a small business. This business has a small number of employees. For various reasons, my client would like to be able to have a copy of all of the emails sent from their employees BCC'd to them. The net effect here would be similar to the access they would have if they hosted their email through Exchange but the business is too small to make this a feasible option. They are currently hosted through GoDaddy. I have not investigated it myself personally but apparently GoDaddy can do something along these lines for all incoming email but not for outgoing email. Is there a way to set up email accounts for a particular domain to where a specified admin user could be copied on all outgoing email?

    Read the article

  • How can I track all emails sent from my users?

    - by Schnapple
    My client runs a small business. This business has a small number of employees. For various reasons, my client would like to be able to have a copy of all of the emails sent from their employees BCC'd to them. The net effect here would be similar to the access they would have if they hosted their email through Exchange but the business is too small to make this a feasible option. They are currently hosted through GoDaddy. I have not investigated it myself personally but apparently GoDaddy can do something along these lines for all incoming email but not for outgoing email. Is there a way to set up email accounts for a particular domain to where a specified admin user could be copied on all outgoing email?

    Read the article

  • subversion: enforce TLS

    - by Daniel Marschall
    Hello, I am running subversion on a Debian Squeeze system with Apache2 and mod_dav for viewing the contents with a webbrowser. I want to enforce the usage of TLS, so that the login data and the SVN contents cannot be read from the connection. I have tried following: <Location /svn> DAV svn SVNParentPath /daten/subversion/ # our access control policy AuthzSVNAccessFile /daten/subversion/access_control # try anonymous access first, resort to real # authentication if necessary. Satisfy Any Require valid-user # how to authenticate a user AuthType Basic AuthName "Subversion repository" AuthUserFile /daten/subversion/.htpasswd # Test SSLRequireSSL RewriteEngine On RewriteCond %{SERVER_PORT} !443 RewriteRule ^svn/(.)$ https://www.viathinksoft.de/svn/$1 [R,L] </Location> at file /etc/apache2/conf.d/subversion.conf Alas, this does not work. There is no redirect and there is still a HTTP request working at /svn/(projectname)/(somefolder) . This SSL-enforce-policy should work for - viewing the contents with webbrowser - retrieve contents with TurtoiseSVN client - committing contents with TurtoiseSVN client Can you please help me? Regards Daniel Marschall

    Read the article

  • How to prevent mod_cluster as single point failure?

    - by Hitesh
    In my configuration i used Apche+mod_cluster as a front_end(load balancer) server and two JBOSS AS 7.1.0 as backend server. In my configuration on one system i have install Apache+mod_cluster + JBOSS AS 7.1.0 and on other system JBOSS AS 7.1.0. Both JBOSS AS 7.1.0 are in domain mode means in clustering. My problem is that if Apche+mod_cluster crash than client can not access any JBOSS AS . I want to configure Apache+mod_cluster in Master_Slave format, means if one of Apache+mod_cluter down another Apache+mod_cluster become active and pass client(browser) request to back-end server in normal form without any interruption. Is there any way to make two Apache+mod_cluster to communicate with each other , means to check health status of each other and if any one get down other Apche+mod_cluster will do it's task ......

    Read the article

  • What port does OpenLink ODBC Driver use?

    - by user36737
    I use Avaya Reporting Services and OpenLink ODBC Drivers for db connection. I know that it uses port 5000 for handshaking but after that I believe it uses an random port for communication. I want to deploy my application and it will communicate with the client's system in their datacenter. They are asking what ports should they open on their firewalls. I can't obviously give them a range above 50,000 that I know OpenLink ODBC Drivers use. Can someone tell me what port should I tell my client to open?

    Read the article

  • stunnel: SSL-to-SSL? (for smtp/imap)

    - by nonot1
    Hello, How can I configure stunnel to accpet SSL connections, and connect then to an SSL port on a different server? Here is my setup: Our ISP's server, "Mail Server", supports smtp/imap over SSL. (Not starttls. Just over ssl.) But, I have a bunch of client machines that will only trust a specific, internal, root certificate. Thus, they can not connect to "Mail Server". For these client machines, I'd like to make a dedicated "Mail Tunnel" host that uses stunnel to listen with an in-house signed SSL certificate, and just forward data to "Mail Server" using a 2nd SSL connection. Can this be done? What would be the specific steps for Ubuntu Server 10.10? (I'm not too familiar with persistent service configuration.) Thank you

    Read the article

  • Missing management console for Symantec Antivirus, Corporate Edition

    - by Alex C.
    I work for a small (80 employees) non-profit organization. We have a Windows network with a domain. The file server is running Server 2003. A vendor recently installed Symantec Antivirus, Corporate Edition, ver. 10.1.8.8000. It used to be the case that when I logged in to the file server as administrator, I would see an icon on the desktop that took me to the management console for the antivirus software. I could check logs, statuses of client machines, set policies, etc. That icon is now missing, and I can't find anything relevant under the "Start" menu. There is an entry for Symantec Antivirus, but that only gives me the panel for the client process, not the overall management console for the whole network. I don't want to call the vendor back to fix this (I'm baffled how we lost the icon in the first place). Can someone offer some guidance?

    Read the article

  • Acceptable placement of the composition root using dependency injection and inversion of control containers

    - by Lumirris
    I've read in several sources including Mark Seemann's 'Ploeh' blog about how the appropriate placement of the composition root of an IoC container is as close as possible to the entry point of an application. In the .NET world, these applications seem to be commonly thought of as Web projects, WPF projects, console applications, things with a typical UI (read: not library projects). Is it really going against this sage advice to place the composition root at the entry point of a library project, when it represents the logical entry point of a group of library projects, and the client of a project group such as this is someone else's work, whose author can't or won't add the composition root to their project (a UI project or yet another library project, even)? I'm familiar with Ninject as an IoC container implementation, but I imagine many others work the same way in that they can scan for a module containing all the necessary binding configurations. This means I could put a binding module in its own library project to compile with my main library project's output, and if the client wanted to change the configuration (an unlikely scenario in my case), they could drop in a replacement dll to replace the library with the binding module. This seems to avoid the most common clients having to deal with dependency injection and composition roots at all, and would make for the cleanest API for the library project group. Yet this seems to fly in the face of conventional wisdom on the issue. Is it just that most of the advice out there makes the assumption that the developer has some coordination with the development of the UI project(s) as well, rather than my case, in which I'm just developing libraries for others to use?

    Read the article

  • Basic AppFabric Service Bus Programming Lifecycle

    - by kaleidoscope
    The tasks required to create an application that access the AppFabric Service Bus are as follows: Create a service namespace. This service namespace contains the resources used by the AppFabric Service Bus to support the application. Define the AppFabric Service Bus contract. A contract specifies the signature of the service, the data it exchanges, and other required inputs, behavior specifications, and object invariants. Implement the contract. To implement a service contract, create a class that implements the interface and specify custom runtime behaviors. Configure the service by specifying endpoint and other behavior information. Build and run the service. Build and run the client application. As with any iterative, service-oriented software development, it may not always be appropriate to follow the preceding steps sequentially, or even start from step 1. For example, if you want to build a client for a pre-existing service, you start at step 5. Or, if you are building a host service that others will use, you can skip step 6. Source: http://msdn.microsoft.com/en-us/library/ee173580.aspx   Sarang, K

    Read the article

  • putty pageant - forget keys after period of inactivity

    - by pQd
    in the environment where windows client computers are used to run putty to connect to multiple linux servers i'm considering moving away from password based authentication and using public/private key pairs with pass-phrases. using ssh-agent would be nice, but at the same time i'd like it to 'forget' the pass-phrases after given period of inactivity. it seems that putty's pageant does not provide such feature; what would you suggest as alternative? solutions that i'm considering: patching pageant code [might be tricky, code is probably quite rusty and project - sadly - stagnant] writing small custom application using GetLastInputInfo and killing pageant if the machine was idle for more than let's say 15 minutes [ yes, there'll be separate policy for locking the desktops as well ] using alternative ssh client and ssh agent. any suggestions? thanks!

    Read the article

  • Best way to replicate servers

    - by Matthew
    I currently have two servers both with linux software RAID1 configurations. They use heartbeat and DRBD to create a shared DRBD device that hosts a a exported NFS directory. The servers run Ubuntu Server with a LXDE GUI and some IP These servers are going to be placed on fishing vessels to act has redundant storage for IP cameras. My boss wants me to figure out the most efficient way to create these servers. We might be looking at pushing out several systems a week. Each configuration will be almost identical besides IP addressing. What would be the best method to automate the configuration process? We are trying to cut down on labor costs to set these up. Imaging and Proceeding are both on my mind right now

    Read the article

  • About the External Graphics Card and CPU usage

    - by Balaji
    We are Rendering 16 live Streams at our client machine through one of our applications and the resolution of the video streams are as 4CIF/MPEG4/25FPS/4000Kbits. The configuration of the client machine is below. HP Desktop Machine: Microsoft Windows XP Intel (R) Core2 Duo CPU E8400 @ 3.00 GHz 2.99 GHz, 1.94 GB of RAM Intel (R) Q45/Q43 Series Express Chipset (Inbuild) The CPU usage of the machine peaks 99% for 16 streams. After some discussion, we had decided to install external graphics card to reduce the CPU usage. So that, we have tried following graphics cards. NVIDIA Quadro NVS 440 - 128 MB Radeon HD 4350 - 512 MB GDDR2 Redeon HD 4350 - 1GB DDR2 ASUS EAH 4350 Silent 1GB DDR2 But the performance wise there has been no difference - even a drop in performance. So, what is the purpose of these external graphics cards? Really it will reduce the CPU usage? What parameters have to check, if we want to reduce the CPU usage?

    Read the article

  • Active node stops resources when pasive node is shutdown

    - by Wakaru44
    2 nodes, active/pasive. 2 resources, a virtual ip, openLdap, and the nfs mount where openldap saves the data. When both nodes are up, things worked fine. You could move resources away and put the active in stanby. But when i rebooted the passive node, ( with the resources in the active node), and the passive node loses conectivity, all the resources in the active where stopped by pacemaker. I'm reading the documentation right now, but I just need a little quick tip to figure what could be hapenning here. Im using: corosync pacemaker RHEL 6

    Read the article

  • A Huge Opportunity in Small Things

    - by Tori Wieldt
    Addressing the strong demand for Java in the embedded market, Oracle is hosting a new Java Embedded @ JavaOne event in San Francisco October 3-4. The event allows decision makers to attend the Java Embedded @ JavaOne business-focused program, while their IT/development staff can attend the technically-focused JavaOne conference. [Obligatory comment about suits & ties vs. jeans & T-shirts removed.] The two-day event includes keynotes, sessions and demonstrations. In his keynote this morning, Judson Althoff, Senior Vice President of Worldwide Alliances and Channels and Embedded Sales, Oracle explained  Devices are all around us - on 24x7, connected all the time. The explosion of devices is the next IT revolution. Java is the right solution for this space. Java embedded solutions provide a framework to  provision, manage, and secure devices.  Java embedded solutions also provide the ability to aggregate, process and analyze multitude of data.  Java is one platform to program them all. Terrance Barr, Java Evangelist and Java ME expert is enthusiastic about the huge opportunity, "It's the right time and right place for Java Embedded," he said, "Oracle is looking for partners who want to take advantage of this next wave in IT." The Embedded space continues to heat up. Today, Cinterion launched the EHS5, an ultra compact, high-speed M2M communication module providing secure wireless connectivity for a wide variety of industrial applications. Last week, Oracle announced Oracle Java ME Embedded 3.2, a complete client Java runtime Optimized for resource-constrained, connected, embedded systems, Oracle Java Wireless Client 3.2, Oracle Java ME Software Development Kit (SDK) 3.2, and Oracle Java Embedded Suite 7.0 for larger embedded devices. There is a huge opportunity in small things. 

    Read the article

  • About online game servers and how to handle data

    - by TreantBG
    So my question isn't about what technology to use or how to do this or that, but a more general question. I'm currently developing a action third person shooter. With elements of RPG - weapon,armor upgrades and items. Players will be able to create new games or join old ones. So my question is how to create the game server that players will play in. I have two ideas on my mind. The player who made the game is the server. All data passes trough him and he send this data to the server updating the database of the players with their XP points kills/deaths score and other. Or my host machine is the server, the player who made the game just will open new instance on my host and will be like client. And all players send their input data to the host, the host updates the game and send response back to client for any new changes like where is the enemy and other. And if i choose option 1 is there a chance the host to change the game content and manipulate the game results? (I think there is but i'm not sure) And if i choose option 2 isn't that raising the response time and potentially the game lag? or maybe there is another option?

    Read the article

  • DHCPOFFER delay VLAN

    - by john883
    I have configured two VLANs [ 15 and 16 ] and a trunkport on a Cisco Catalyst 2960. The trunkport is connected to eth2 on a Linux server The server is configured to support VLAN's and the interfaces eth2.15 eth2.16 is configured with ip addresses on two different subnet. dhcp3-server is running on the same server and hands out IP-addresses to the VLANs. When connecting a client to a port that is configured in ex. VLAN 15 and requesting a IP-address, i experience a long delay before recieving a DHCPOFFER, around 30 seconds or so, the client needs to send a DHCPDISCOVER about five times but will always recieve a DHCPOFFER. Any suggestion why this delay is happening?

    Read the article

< Previous Page | 300 301 302 303 304 305 306 307 308 309 310 311  | Next Page >