Search Results

Search found 87589 results on 3504 pages for 'veritas cluster server'.

Page 629/3504 | < Previous Page | 625 626 627 628 629 630 631 632 633 634 635 636  | Next Page >

  • How does one make sure or even guarantee server time are sync correctly between dozens of servers across multiple datacenter on different location?

    - by forestclown
    Currently our web applications contain a logic to check if the data sent to the web server is expired or not by comparing the timestamp of the data with the date/time of the server. Everything goes will, until some dude from data center accidentally modify one of the web server date/time and causes some disruptions in our web services. My managers are of course not happy with this, and said we shouldn't use timestamp to check expiry in the first place...anyway.... Network Time Protocol is implemented, because of data centers are spread across different continents so we have one NTP server in each data center. The servers within the data center will have cron jobs to check against the time with their NTP server from the same data center. If time is out of sync it will auto update the server date/time. But then with our managers not happy with it, and think it could still easily causes the same problem. e.g. what if someone accidentally modify the NTP date/time? what if all the NTP servers are out of sync with each other? which NTP servers we can really trust? and blah blah.. So my questions are: What are the current practice to sync date/time between servers across multiple data centers or locations? How does one manages time stamp between web apps? e.g. Server A send data (contain timestamp of Server A) to Server B (compare timestamp between Server B and the timestamp from the data to see if it has expired or not. This is to avoid HTTP replay) Should we really not use timestamp check? Thanks & Best Regards

    Read the article

  • Online Windows Server monitoring

    - by Khash
    To check a website's availability there is Pingdom (and many others). I'm looking for a similar service (online/web-based, easy to use with notifications) that monitors servers a bit more in detail. Things like Disk Space, Windows Services running, etc.... I am happy to install an agent on the box to facilitate that, but don't want to run the monitoring server as well.

    Read the article

  • Upgrade from Etch to Lenny, server no longer reboots

    - by Jack
    Hi, I just upgraded my (remote) webserver to Lenny from Etch. and no real issue occurs except that I cannot reboot that server. (I can only start in a rescue mode) I recall that a MBR change performed and I am wondering if that could be the reason for this failure. Would someone know how I can revert this MBR change ? I had switched back to my previous linux kernel, still cannot reboot. I would appreciate your help. thanks Jack

    Read the article

  • pg_dump: Error message from server: ERROR: missing chunk number 0 for toast value 43712886 in pg_to

    - by Kirill Novozhilov
    After $ pg_dumpall -U postgres -f /tmp/pgall.sql I see following: pg_dump: SQL command failed pg_dump: Error message from server: ERROR: missing chunk number 0 for toast value 43712886 in pg_toast_16418 pg_dump: The command was: COPY public.page_parts (id, name, filter_id, content, page_id) TO stdout; pg_dumpall: pg_dump failed on database "radiant", exiting I haven't earlier backups. How can I fix it? Thanks in advance.

    Read the article

  • Access denied on virtual machine

    - by newbie
    I have a friend on the same network as me hosting a virtual machine using vmware server.The guest OS is Windows 7 and Windows Server 2003.He granted me administrative priveleges for the machine but when i try to add a user to that VM it says access denied. Moreover when i try to start the VM remotely using the command vmrun -T server -h https://computer-name.domain.com:8333/sdk -u domain/user -p passwd start "[standard] Win2k3Client1/Win2k3Client1.vmx" This shows unsufficient priveleges on host operating system but i am able to login with same username password on VMWare Web Access

    Read the article

  • how to connect to server continuously using an bad internet connetion

    - by Nikhil
    I have a bad Internet connection, it disconnects frequently and on reconnect, I'm assigned a different IP address by the ISP. The problem is that I connect to a remote VPS (Ubuntu), and when Internet connection is disrupted n reconnected, I can no longer do anything on the terminal. I have to restart the terminal and re-initiate the connection. Is there a way I can have persistent connection with server.

    Read the article

  • Xen Server Guest Network configuration?

    - by GeekTantra
    I am trying to install Ubuntu Server as a domU (Guest) on Citrix XenServer. Everything works fine but I am not able to access the internet from the guest. Is there any specific thing that I need to do to configure the network for the Guest. I have two NICs, 1 has an external IP and the other has an internal IP.

    Read the article

  • Best way to provide redundant switching/links to server

    - by Myles Gray
    We have 3x ESX hosts and 2x SANS that we wish to move to a redundant 10G networking infrastructure. We have 4x Dell PowerConnect 8024F's to provide our backbone and are configured as so (only core switches relevant to this question): So the questions are: 1) Do the interconnects between the 4x 8024F's need to be LAG'd or just STP'd 2) As the NICs on the servers are split across 2 switches, does any special configuration need to be done here or on the switches? 3) If a link or switch fails will the switches automatically find a new path to the Server/SAN?

    Read the article

  • Slow SSD Read on Server 2008 SP2

    - by Ruben L.
    Im using a Intel X25-M 80G with the latest firmware. on XP /w using IDE and WIN7 using AHCI I get read speeds up to 250MB/S. But when running it with Server 2008 SP1 or SP2 on AHCI, I get read speeds around 180MB/S. Ive updated drivers for 2008, tested with writecache on/off. Any input would be appreciated. thanks!

    Read the article

  • Server Density or Scout-like Services for Windows?

    - by Michael Gorsuch
    Hi all. I've been very impressed with agent-based / SaaS-like monitoring services such as Scout and Server Density. They both provide very nice trending, alerting, and reporting capabilities for my *nix boxes, and I need to find something similar for Windows. Can anyone out there recommend such a tool? It is important that this tool is a centralized service in line with the aforementioned products, as the deployment environment is owned by a lean team with limited resources. Thanks!

    Read the article

  • Internal server error,godaddy problem?

    - by user303832
    I added subdomain,create a new folder,than put some files in folder,when I try to access my new subdomain I get this message 500 - Internal server error. There is a problem with the resource you are looking for, and it cannot be displayed. Can someone help me with this please?It is linux hosting with mysql database.

    Read the article

  • Securing ColdFusion for internet facing server

    - by Goyuix
    What do I need to do to tighten down a ColdFusion server for internet facing apps? The only thing that specifically came to mind was to restrict the CFIDE and JRunScripts directories to a local subnet. Are there settings in the administrator I can tweak to make the applications more secure?

    Read the article

  • What's the best way to do user profile/folder redirect/home directory archiving?

    - by tpederson
    My company is in dire need of a redesign around how we handle user account administration. I've been tasked with automating the process. The end goal is to have the whole works triggered by the business, and IT only looking in when there's an error reported. The interim phase is going to be semi-manual. That is a level 2 tech inputs the user's info and supervises the process. The current hurdle I'm facing is user profile archiving. Our security team requires us to archive the profile directories for any terminated user for 60 days in case the legal team requires access to their files. Our AD is as much a mess as everything else, so there are some users with home directories and some with profiles. Anyone who has a profile dir in AD also has a good deal of their profile redirected to our file servers over DFS. In order to complete the process manually you find the user in AD, disable them, find their home/profile dir, go there and take ownership, create an archive folder, move all their files over, then delete the old dir. Some users have many many gigs of nonsense and this can take quite some time. Even automated the process would not be a quick one. I'm thinking that I need to have a client side C# GUI for the quick stuff and some server side batch script or console app to offload this long running process. I have a batch script that works decently using takeown and robocopy, but I wonder if a C# console app would do a better job. So, my question at long last is, what do you think is the best way to handle this? I can't imagine this is a unique problem, how do other admins get this done? The last place I worked was easily 10x larger than the place I'm in now. If we would have been doing this manual crap there, they'd have needed a team of at least 30 full time workers to keep up. I have decent skills in C#.net and batch scripting, but am a quick study and I have used most every language once or twice. Thank you for reading this and I look forward to seeing what imaginative solutions you all can come up with.

    Read the article

  • Exposing server uptime a potential security vulnerability?

    - by Edward Mazur
    I run a browser-based game and as part of a page with a bunch of game statistics, I have the server uptime listed. It's currently at 177 days and so someone mentioned to me that I shouldn't do this because a long uptime indicates the kernel is old and therefore missing security updates. This certainly sounds logical, but I searched around and couldn't find any evidence to support it. So I'm just wondering, is this indeed something I should not have exposed?

    Read the article

  • networked storage for a research group, 10-100 TB

    - by Marc
    this is related to this post: http://serverfault.com/questions/80854/scalable-24-tb-nas-for-research-department but perhaps a little more general. Background: We're a research lab of around 10 people who do a lot of experiments that involve taking pictures at one of several lab setups and then analyzing it an one of several lab computers. Each experiment may produce 2 or 3 GB of data, and we are generating data at the rate of about 10 TB/year. Right now, we are storing the data on a 6-bay netgear readynas pro, but even with 2 TB drive, this only gives us 10 TB of storage. Also, right now we are not backing up at all. Our short term backup plan is to get a second readynas, put it in a different building and mirror the one drive onto the other. Obviously, this is somewhat non-ideal. Our options: 1) We can pay our university $400/ TB /year for "backed up" online storage. We trust them more than we trust us, but not a whole lot. 2) We can continue to buy small NASs and mirror them between offices. One limit, although stupid, is that we don't have an unlimited number of ethernet jacks. 3) We can try to implement our own data storage solution, which is why I'm asking you guys. One thing to consider is that we're a very transient population and none of us are network administration experts. I will probably be here only another year or so, and graduate students, who are here the longest, have a 5-6 year time scale. So nothing can require expert oversight. Our data transfer rates are low - most of the data will just sit on the server waiting for someone to look at it once or twice - so we don't need a really high speed system. Given these contraints, can someone recommend a fairly low-cost, scalable, more or less turn key shared data storage system with backup in a separate physical location. Does such a thing exist or should we just pay the university to take care of it for us? As a second question, our professor just got tenure and is putting together a budget. Here the goal is to ask for as much as you can and hope you get a fraction of it. So the same question, minus the low-cost. Without budget constraints, can you recommend a scalable turn-key backed up storage system. Thanks

    Read the article

  • Looking for good Open Source email server

    - by rockinthesixstring
    I'm currently running MailEnable as my email server, it's ok but lacking. I'm looking for a better alternative and am wondering about Open Source. I'm a huge fan of the Smarter Mail, and will pay if it's the best I can get... But I have been learning that there is some good stuff out there in the Open Source community (IE: I've started using PFSense and I love it). Does anyone have any suggestions?

    Read the article

< Previous Page | 625 626 627 628 629 630 631 632 633 634 635 636  | Next Page >