Search Results

Search found 2115 results on 85 pages for 'chum of chance'.

Page 28/85 | < Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >

  • Shared Excel WorkBook is locked by another user

    - by Simone
    I’ve been trying everything; this is the last chance I have. I moved folders and files from an old Windows Server 2003 File Server to a new FS (Win Server 2008 R2) with DFS and ABE enabled. Now, a specific Shared Excel file is driving me crazy, out of a sudden, lots of times per day, users are getting the following error while opening that file: Filename.xlsx is locked for editing by ‘another user’. Open ‘Read-Only’ or, click ‘Notify’ to open.. I’ve already followed this, with no joy: http://blogs.technet.com/b/the_microsoft_excel_support_team_blog/archive/2012/05/14/the-definitive-locked-file-post.aspx In any case, I strongly think this is not client-related, since it never gave that problem in the past with Windows Server 2003. I’ve found and followed many other solutions, nothing. The users are all utilizing Office 2010 on Windows 7 machines, besides a few users who are still on Windows XP machines. I appreciate any help, thank you!

    Read the article

  • Shared Excel WorkBook is locked by another user

    - by Simone
    I’ve been trying everything; this is the last chance I have. I moved folders and files from an old Windows Server 2003 File Server to a new FS (Win Server 2008 R2) with DFS and ABE enabled. Now, a specific Shared Excel file is driving me crazy, out of a sudden, lots of times per day, users are getting the following error while opening that file: Filename.xlsx is locked for editing by ‘another user’. Open ‘Read-Only’ or, click ‘Notify’ to open.. I’ve already followed this, with no joy: http://blogs.technet.com/b/the_microsoft_excel_support_team_blog/archive/2012/05/14/the-definitive-locked-file-post.aspx In any case, I strongly think this is not client-related, since it never gave that problem in the past with Windows Server 2003. I’ve found and followed many other solutions, nothing. The users are all utilizing Office 2010 on Windows 7 machines, besides a few users who are still on Windows XP machines. I appreciate any help, thank you!

    Read the article

  • Fatal error: Incompatible file format: The encoded file has format major ID 1, whereas the Loader expects 4 in ... on line 0

    - by Eugene
    I am using Ubuntu 10.04 and for some time I had to keep a downgraded PHP 5.2 package because I need to run Zend encrypted scripts. Recently I noticed that Zend released beta version of their loader (http://forums.zend.com/viewtopic.php?f=57&t=1365&start=80#p22073) so I updated to the native PHP 5.3 package, downloaded the .so file, added this to php.ini ;zend_extension=/etc/php5/ZendOptimizer.so zend_extension=/etc/php5/ZendGuardLoader.so zend_loader.enable=1 zend_loader.disable_licensing=0 zend_loader.obfuscation_level_support=3 and restarted the server. Now I am getting this error: Fatal error: Incompatible file format: The encoded file has format major ID 1, whereas the Loader expects 4 in ... on line 0 Do you by chance know an easy fix for this? Or should I downgrade back and wait till when they release something more stable?

    Read the article

  • Reconfigure RAID on Dell PowerEdge T710

    - by Stefano Borini
    I have a Dell PowerEdge T710 under my feet at this very moment, with RedHat Enterprise Server 5.3. I have 6 1TB disks and two 500GB. parted reports two devices, one 500 GB and the other 4 TB. So I assume the RAID has been setup as mirror for two disks, and I assume as RAID 5 the remaining ones. I say "I assume" because it does not make sense. Having 6 disks in RAID 5, I should obtain a total space of 5 TB, not 4TB. It's not even RAID 10: I would end up with a 3 TB unit. How can I check and eventually modify the RAID array definition? In the Fujitsu Siemens I played with some time ago, at boot I had the chance to enter the controller BIOS, but here I don't see a clear way to perform this operation.

    Read the article

  • Kernel NTFS driver vs NTFS-3G

    - by Jack
    A more comprehensive phrased question since I lost access to the other one. I would ask that the other one be deleted, not this one, as it should not have been migrated in the first place. There are currently two NTFS drivers available for Linux. The NTFS driver included in the kernel, and the userspace NTFS-3G driver that makes use of FUSE. By all accounts, NTFS-3G works perfectly. My question then, is if the NTFS filesystem has been successfully reverse engineered, why have the kernel NTFS team not implemented the changes in their driver? At the moment it is still marked as experimental, and there is a good chance it will destroy your data. Note: This has absolutely nothing to do with distributions...

    Read the article

  • Firewall behind firewall

    - by Makach
    I've recently changed jobs and I've been set up with a new workstation. On all previous places where I've been working they've had some sort of local firewall installed on each and every workstation - but here I've been told not to activate it because it is not necessary since we're already behind a HW Firewall. To me this seem a bit naïve, but I cannot emphasise it. I always thought a local firewall was good practice, ie. if something managed to come through the hw firewall there might be a slight chance other computers on the lan would block the internal threath. We got free access to internet and we got a virus checker installed.

    Read the article

  • We have no SW Firewall behind our office HW firewall, admin says its not req'd

    - by Makach
    I've recently changed jobs and I've been set up with a new workstation. On all previous places where I've been working they've had some sort of local firewall installed on each and every workstation - but here I've been told not to activate it because it is not necessary since we're already behind a HW Firewall. To me this seem a bit naïve, but I cannot emphasise it. I always thought a local firewall was good practice, ie. if something managed to come through the hw firewall there might be a slight chance other computers on the lan would block the internal threath. We got free access to internet and we got a virus checker installed.

    Read the article

  • Running a scheduled task as SYSTEM with console window open

    - by raoulsson
    I am auto creating scheduled tasks with this line within a batch windows script: schtasks /Create /RU SYSTEM /RP SYSTEM /TN startup-task-%%i /TR %SPEEDWAY_DIR%\%TARGET_DIR%%%i\%STARTUPFILE% /SC HOURLY /MO 1 /ST 17:%%i1:00 I wanted to avoid using specific user credentials and thus decided to use SYSTEM. Now, when checking in the taskmanagers process list or, even better, directly with the C:\> schtasks command itself, all is working well, the tasks are running as intended. However in this particular case I would like to have an open console window where I can see the log flying by. I know I could use C:\> tail -f thelogfile.log if I installed e.g. cygwin (on all machines) or some proprietary tools like Baretail on Windows. But since I only switch to these machines in case of trouble, I would prefer to start the scheduled task in such a way that every user immediately sees the log. Any chance? Thanks!

    Read the article

  • NFS and KVM. Slow Speed

    - by Javier Martinez
    I have a KVM virtualization in Debian with 2 guests (Debian and Windows 2008). I want to have a 'mount point' shared that can be accessed by the 3 system (host and 2 guests) at the same time. So the only thing that I found was a NFS/SMB network storage. I picked NFS Due to my Ethernet network (10/100), the speed average that I get between accessing/transfering files between the 3 system is always 8~10MB/s. The point is if is there any chance of get a boost system for sharing files between 3 system (at the same time) without wasting the speed of my SATA disks. I mean, without the Ethernet limitation of 10 MB/s

    Read the article

  • Get rid of Vista security warning

    - by Ken
    I found this question. The question exactly matches my problem, but the solution doesn't work. In the Properties window, I see "Security: This file came from another computer and might be blocked to help protect this computer. ((Unblock))". When I click Unblock and Apply, the Security section disappears. But when I go to run it again, I still get the security warning. If I right-click and choose Properties on the exact same thing, the Security section is back, offering me the chance to Unblock it again. So unblock seems exactly as useless as the "Always ask" checkbox. Anyone seen this before? How do you really Unblock an app that Vista doesn't want to let you Unblock?

    Read the article

  • Keyboard Shortcuts in Win 7 without the CTRL + ALT

    - by Carlos
    I am knew to this site and don't know if I'm doing this correctly. I've been asked to edit my original post so I deleted my original post and starting over. I don't know why it's so hard for everyone to understand what I'm trying to do. You guys are all geniuses when it comes to computers and I'm just starting out. I started out trying to use a shortcut to display the LOCAL AREA CONNECTION window on my desktop by creating a shortcut and assigning it CTRL + , (comma). Windows didn't like that so it added ALT which ended up being CTRL + ALT + ,. Since I couldn't figure out a way to eliminate ALT as part of the shortkey keys, I am now trying a different strategy and it's not working. my latest attempt is to run the following command; ^,:: Run, explorer:: {BA126ADB-2166-11D1-B1D0-00805FC1270E} Can someone please tell me what I'm doing wrong? I'm trying, just give me a chance. Thanks, Carlos

    Read the article

  • single point of failure in IIS Web Farm Framework setting?

    - by aamir sajjad
    ASP.NET WEB API Windows Server 2008 R2/IIS 7.5/Web Farm Framework 2.5 I am planning to deploy application across 4 web servers. Should i use shared content/configuration using DFS among web servers for web farm scenario? Second option is to use Web Farm Framework for deployment. Furthermore, is there chance of single point of failure in WFF? for example what if primary server goes down. which option would be better? pros and cons of each of the above. I appreciate your response.

    Read the article

  • The User Profile Service

    - by Daryl Gill
    I tried logging into my computer after an "ungraceful" shutdown due to a power cut.. I got prompted with a message stating the user profile service failed the login. After reading a k/b on Microsoft I managed to re-login to the "corrupt" account by making registery edits.. But the next step (After having the ability to logon) is a major concern for me.. Would the guys at superuser suggest: 1) Carry on using said account 2) Re-install Windows to eliminate the possible chance of a reoccouring problem 3) Migrate all user data over to a new account and with that being said above. I had to issue the command line: net user administrator /active:yes to activate the hidden adminstrator account, would it be recommended/secure to disable the account or leave it active?

    Read the article

  • Requiring SSH-key Login Via PAM From Specific IP Ranges

    - by Sean M
    I need to be able to access my server (Ubuntu 8.04 LTS) from remote sites, but I'd like to worry a bit less about password complexity. Thus, I'd like to require that SSH keys be used for login instead of name/password. However, I still have a lot to learn about security, and having already badly broken a test box when I was trying to set this up, I'm acutely aware of the chance of screwing myself while trying to accomplish this. So I have a second goal: I'd like to require that certain IP ranges (e.g. 10.0.0.0/8) may log in with name/password, but everyone else must use an SSH key to log in. How can I satisfy both of these goals? There already exists a very similar question here, but I can't quite figure out how to get to what I want from that information. Current tactic: reading through the PAM documentation (pam_access looks promising) and looking at /etc/ssh/sshd_config.

    Read the article

  • Does the OS X DHCP client support classless static routes (RFC3442)?

    - by cense
    Simply put: does Mac OS X support RFC3442? RFC3442 implements DHCP codes (121 and 249) to provide classless static routes to DHCP clients. I can only seem to find rather old information circa 2005-2008 which implies that OS X does not support RFC3442. Any chance this has changed recently and I'm just doing something wrong while setting it up? My anecdotal testing seems to confirm that OS X does not support RFC3442 but I'm looking for confirmation. If OS X still doesn't support RFC3442, what is an alternative method of pushing static routes to OS X clients?

    Read the article

  • Oracle on NFS vmdk beats native NFS!?

    - by fletch00
    Hi, my colleagues are pursuing this with Netapp and Oracle - but I thought I'd post here on the off chance someone else has seen this We have a RedHat 5 VM (fully up2date) running Oracle 11i with data disks mounted via the VM's linux kernel NFS using Oracle's recommended mount options and the performance is very inconsistent (Querys that should take < 2 seconds sometimes take 60 seconds) Funny thing is we can run the same queries perfectly consistently < 2 seconds on a VMDK residing on SAME NetApp NFS datastore! Makes me wish Oracle and NetApp collaborated as closely as VMware and NetApp did on the Virtual Storage Console we used to perfectly set the NFS options and keep them in compliance... We have tried a few Linux NFS options others have posted and not seen improvement so far. We are now creating VMDK's for the VM to replace the Linux NFS mounted and workaround the issue as our developers need consistent performance ASAP.

    Read the article

  • TIME_WAIT connections not being cleaned up after timeout period expires

    - by Mark Dawson
    I am stress testing one of my servers by hitting it with a constant stream of new network connections, the tcp_fin_timeout is set to 60, so if I send a constant stream of something like 100 requests per second, I would expect to see a rolling average of 6000 (60 * 100) connections in a TIME_WAIT state, this is happening, but looking in netstat (using -o) to see the timers, I see connections like: TIME_WAIT timewait (0.00/0/0) where their timeout has expired but the connection is still hanging around, I then eventually run out of connections. Anyone know why these connections don't get cleaned up? If I stop creating new connections they do eventually disappear but while I am constantly creating new connections they don't, seems like the kernel isn't getting chance to clean them up? Is there some other config options I need to set to remove the connections as soon as they have expired? The server is running Ubuntu and my web server is nginx. Also it has iptables with connection tracking, not sure if that would cause these TIME_WAIT connections to live on. Thanks Mark.

    Read the article

  • Which IMAP flags are reliably supported across most mail servers?

    - by Ben Butler-Cole
    I am writing an application which reacts to emails sent to a mailbox. It retrieves the emails via IMAP. It will be deployed to a number of systems where I do not control the mail server configuration. I would like to use IMAP flags to indicate which messages have been handled. Are the system flags sufficiently widely supported that I can reasonably depend on them in my application? Are user-defined flags sufficiently widely supported? (If the answer is "ha ha, not a chance", then I shall use folders instead.) Thanks -Ben

    Read the article

  • Using DD for disk cloning

    - by roe
    There's been a number of questions regarding disk cloning tools (yes, I did search for it) :) and dd has been suggested at least once. I've already considered using dd myself, mainly because ease of use, and that it's readily available on pretty much all run of the mill bootable linux distributions. To my question, what is the best way to use dd for cloning a disk. I don't have much time to go around, and no test-hardware to play with, so I need to be prepared, and it's pretty much a one-shoot chance to get it done. I did a quick google search, and the first result was an apparent failed attempt. Is there anything I need to do after using dd, i.e. is there anything that CAN'T be read using dd? Thanks!

    Read the article

  • Is it common to have portable hard drive not recognize?

    - by lamwaiman1988
    I have a bad hard drive which could not be recognized by windows. If windows can recognize it, it could just be pure luck. There are better chance of it being recognize by plugging it at the usb back of the machine, rather than the front one. More interestingly, I have a usb cable which fork itself at one side ( i.e it has 2 heads at one side ). I can plug both heads to the machine and connect the portable hard drive on the other end of the usb cable? Can I get more voltage by doing this? I am guessing the front usb port doesn't have enough power. I am wondering if I buy a new portable usb hard drive, would it have easy read by plugging at the front usb port? I am not sure if I screw up the jumper connecting the front usb ports, but it is okay for other device such as usb stick and mouse.

    Read the article

  • Running a scheduled task as SYSTEM with console window open

    - by raoulsson
    I am auto creating scheduled tasks with this line within a batch windows script: schtasks /Create /RU SYSTEM /RP SYSTEM /TN startup-task-%%i /TR %SPEEDWAY_DIR%\%TARGET_DIR%%%i\%STARTUPFILE% /SC HOURLY /MO 1 /ST 17:%%i1:00 I wanted to avoid using specific user credentials and thus decided to use SYSTEM. Now, when checking in the taskmanagers process list or, even better, directly with the C:\> schtasks command itself, all is working well, the tasks are running as intended. However in this particular case I would like to have an open console window where I can see the log flying by. I know I could use C:\> tail -f thelogfile.log if I installed e.g. cygwin (on all machines) or some proprietary tools like Baretail on Windows. But since I only switch to these machines in case of trouble, I would prefer to start the scheduled task in such a way that every user immediately sees the log. Any chance? Thanks!

    Read the article

  • Find which php scripts cause high CPU with php-cgi

    - by Oli
    Background: I maintain a server for a client who has half a dozen Wordpress sites on. They all have the W3 Total Cache plugin installed and eAcellerator is installed (might be APC). All the PHP sites run through a single batch of fastcgi php-cgi processes (it's actually php-fpm but I'm not sure if that makes a difference). Problem: php-cgi's CPU usage is quite high. Not terminally high but high enough to raise an eyebrow. The client wants to add more sites in the future and I want to avoid becoming CPU limited if I can help it. Question: Is there any way I can find the scripts or even just requests that are causing the high CPU. I realise I might not be able to do anything with the results but it would give me a chance.

    Read the article

  • How to use WeirdX applet instead of X11 for remote connections in Mac OS X?

    - by penyuan
    When I am away, I've got an Apple PowerBook set up to use X11 forwarding with SSH to connect to a remote iMac client in my room for server administration. Whenever I start X client programs on the client iMac, X11 would start on my PowerBook, which is OK, but... Question: Is it possible to change my settings so I can use the WeirdX (a web-based java X server) applet to take care of the X client programs instead of Mac OS X's default X11? Is there a config file I can edit to chance the $DISPLAY variable, and what should I set it to? Is this even the issue? Thank you very much.

    Read the article

  • For virtual machines, when SMP is available on the host, should guest also have SMP setup?

    - by supercheetah
    I'm trying to find out the best "bang for my buck" so to speak in regards to virtual machines, and SMP. I have an Intel Core 2 Duo, which of course has two cores and the VT extensions, and I'm running Ubuntu Linux (host) on it with VirtualBox, which has Windows Vista (guest). Currently I've got the guest machine setup for two processors to give Windows a chance to manage its own parallelism, but I'm not certain that it's any faster. I've tried it with just one processor, but it's hard to tell if it's any better. Any thoughts? Should the guest have two processors setup?

    Read the article

  • How to use WeirdX applet instead of X11 for remote connections in Mac OS X?

    - by penyuan
    When I am away, I've got an Apple PowerBook set up to use X11 forwarding with SSH to connect to a remote iMac in my room. Whenever I start X client programs on the client iMac, X11 would start on my PowerBook, which is OK. Question: Is it possible to change my settings so that the WeirdX applet I have on my PowerBook will take care of the X client programs instead of the default X11? Is there a config file I can edit to chance the $DISPLAY variable, and what should I set it to? Is this even the issue? Thank you very much.

    Read the article

< Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >