Search Results

Search found 1915 results on 77 pages for 'identical'.

Page 25/77 | < Previous Page | 21 22 23 24 25 26 27 28 29 30 31 32  | Next Page >

  • Strange strace and setuid behaviour: permission denied under strace, but not running normally.

    - by Autopulated
    This is related to this question. I have a script (fix-permissions.sh) that fixes some file permissions: #! /bin/bash sudo chown -R person:group /path/ sudo chmod -R g+rw /path/ And a small c program to run this, which is setuided: #include "sys/types.h" #include "unistd.h" int main(){ setuid(geteuid()); return system("/path/fix-permissions.sh"); } Directory: -rwsr-xr-x 1 root root 7228 Feb 19 17:33 fix-permissions -rwx--x--x 1 root root 112 Feb 19 13:38 fix-permissions.sh If I do this, everything seems fine, and the permissions do get correctly fixed: james $ sudo su someone-else someone-else $ ./fix-permissions but if I use strace, I get: someone-else $ strace ./fix-permissions /bin/bash: /path/fix-permissions.sh: Permission denied It's interesting to note that I get the same permission denied error with an identical setup (permissions, c program), but a different script, even when not using strace. Is this some kind of heureustic magic behaviour in setuid that I'm uncovering? How should I figure out what's going on? System is Ubuntu 10.04.2 LTS, Linux 2.6.32.26-kvm-i386-20101122 #1 SMP

    Read the article

  • Most suitable high availability solution

    - by Alex Bagnolini
    My company is hosting a website in a server with IIS, SQL Server and a 3rd party windows service (written in C#, source code available for amendments). We bought a new identical server, composed by: 1x Quad Core, 12GB RAM, 4x160GB SATA Raid 5, Windows Server 2008 R2 Datacenter, Public IP. We aim to put all webpages and the 3rd party windows service in an high-availability state. After some lab-testing on how to configure Failover Clustering and Hyper-V, we have deep doubts on what the "best" solution would be, by "best" meaning maintainable and able to correctly handle a physical server failure. Any suggestion on how we should configure the two servers? We don't need all the configuration's step, just an hint on the right direction to follow.

    Read the article

  • Printing to shared printers across VPN

    - by CYMR0
    I have a program that prints labels at five remote sites. Two sites, aren't working, but the rest are with an identical (as far as I can tell) setup. Using Wireshark, I have determined that the handshaking all goes well, but after the "Open Print File Response" the packet that is sent from the server, doesn't reach the client. But I'm a bit at a loss as to where I go from here. I know the port the packet was sent on (445) isn't being blocked, the RST packet gets sent on the same port and that gets there fine. It's also weird that the three out of five sites are working fine. This has been up and running for years without issue, all that we have changed is our connectivity (from DSL to bonded DSL). But this traffic is over a VPN - so it can't be the ISP interfering either can it? I'm totally stuck, and any help would be much appreciated. Thanks!

    Read the article

  • AuthInfoRequired cups overwrites

    - by mooscape
    My problem is basically identical to the following: http://bbs.archlinux.org/viewtopic.php?id=61826 Put simply, I have a machine in ubuntu trying to connect to another ubuntu machine via a network in order to use the printer attached. There is no problem printing until I restart the guest machine. Immediately it overwrites the printers.conf file (under /etc/cups/printers.conf). It always adds the same line: AuthInfoRequired username,password I stop cups and change it to *#*AuthInfoRequired username,password to comment out the command. Start cups. Works great 'til the next shutdown. Then it gets overwritten again. Googling indicates it may be GTK problem and not CUPS, but I have found no permanent solution to date. Any suggestions appreciated ....

    Read the article

  • copy boot-able partition

    - by Dima
    I have an disk image with 3 partitions: first partition (hd0,0) is boot-able with GRUB1 with the following configuration GRUB file: default=0 timeout=5 title Bank A root (hd0,1) chainloader +1 title Bank B root (hd0,2) chainloader +1 The partitions (hd0,1) and (hd0,2) are also boot-able. I'm trying to clone partition (hd0,1) to (hd0,2) by creating device map using kpartx and copying whole partition using dd command. The problem is: after partition cloning, the cloned partition did not boot (but all files are OK). What the wrong? I need both partitions to bee identical (I'm using them for fail-over purposes into embedded device)

    Read the article

  • How to restore a windows 7 system from a secondary drive

    - by Klas Mellbourn
    I have a stationary computer with Windows 7 Ultimate 64 bit. The primary (SSD) hard drive seems to have stopped working completely, it is not even visible in BIOS. The computer has a secondary hard drive (non-SSD, NTFS, 2TB). I have had Windows backup running and saving backups to that secondary drive. I am planning to buy a new SSD drive to replace the faulty one. I want to restore the backup to this new SSD drive. What is the most straightforward way to do this? A step-by-step description would be greatly appreciated. Further information: I have a Windows install DVD and the computer has a DVD-drive. The secondary drive is not bootable, so I cannot currently access it. The new SSD drive will probably not be identical to the original, so it might need different drivers

    Read the article

  • 4 - 7 second delay accessing Mysql across the network

    - by Kristiaan
    Hello, our company has recently purchased a new server with the intention of replacing our aging database server. its a full 64bit 2008 enterprise system, i have got the basic server setup and configured and then installed the 64bit version of mysql on the server, this has then been configured to match where possible our existing server as much as it can. however i have noticed that when it was swapped with the production database server our software systems had an increased delay accessing the mysql database this was anything beween 4 - 7 seconds. i have tried disabling TOE, IPv6 and a few other suggested soultions to this but so far cannot find out where this slowdown is coming from. replacing the server with the production one and the delay goes away. in terms of software and hardware the servers are not very identical at all due to one being windows 2003 std with a 32bit server and the new one being windows 2008 enterprise with a 64bit server. thanks Kris

    Read the article

  • failing to achive tunneling to fresh ubuntu 10.04 server

    - by user65297
    I've just set up a new 10.04 server and can't get the tunneling to work. local machine > ssh -L 9090:localhost:9090 [email protected] login success, but thereafter trying tunnel from local browser, http://127.0.0.1:9090 echo at server terminal: channel 3: open failed: connect failed: Connection refused auth.log sshd[24502]: error: connect_to localhost port 9090: failed. iptables -L Chain INPUT (policy ACCEPT) target prot opt source destination Chain FORWARD (policy ACCEPT) target prot opt source destination Chain OUTPUT (policy ACCEPT) target prot opt source destination Trying 9090 at server (links http://xx.xxx.xx.xx:9090 works) sshd_config is identical to previous 8.04 server, working fine. What's going on? Thankful for any input. Regards, //t

    Read the article

  • Are there pitfalls to using incompatible RAM (frequencies) in motherboards?

    - by osij2is
    I'd like to use 2 x 4GB DDR3 1600 dimms in a motherboard capable of only DDR3 1066. The DDR3 1600 is on sale and the cost is identical to 1066 dimms. It'd be nice to have these faster sticks around should i upgrade the motherboard. I assume the RAM can under clock itself or be changed in the BIOS. While obviously it's less than ideal situation, I don't know if there are other unintended consequences in terms of stability, performance and longevity of the board and said RAM. Am I doing any damage to the memory controller or RAM? I've always bought RAM at the max speed specified for the motherboard and I've never gone over so I'm not sure if there any caveats to this at all. Edit: I intend to use the RAM in pairs. I know that mixing RAM speeds is just a bad idea.

    Read the article

  • Calculate minimum ext3 partition size for certain amount of data

    - by Daniel Beck
    These following ext3 partitions contain identical data. As we can see, the larger the partition size, the more space is required for the same files: Filesystem 1K-blocks Used Available Use% Mounted on /dev/loop11 3965777 561064 3199964 15% [...] /dev/loop19 573029 543843 29186 95% [...] Filesystem Size Used Avail Use% Mounted on /dev/loop11 3.8G 548M 3.1G 15% [...] /dev/loop19 560M 532M 29M 95% [...] Filesystem Inodes IUsed IFree IUse% Mounted on /dev/loop11 1024000 1656 1022344 1% [...] /dev/loop19 1024000 1656 1022344 1% [...] I start with a partition of fixed size that possibly wasted a lot of space and I want to create a partition that is able to hold that data but with (almost) minimal size. How can I reliably calculate that minimal partition size needed for storing a certain amount of data? The amount of data changes over time, and I need to automate these calculations.

    Read the article

  • Windows - how to clone network connections entry to another machine

    - by tomash
    Let's assume we have two Windows XP computers with identical hardware configuration, including GPRS modem connected to COM1 serial port. On the first PC we add new modem device, set up gprs connection on this port in "Network connections" folder. How avoid clicking the same procedure on second computer? Detailed questions: Where are the entries from "Network connections" stored? File, registry? Is there a possibility to clone such entry between PCs? Question as above but regarding modem entry in device manager - can it be exported from one pc and imported on second?

    Read the article

  • Can one build a single disk RAID 1 array?

    - by Core Xii
    I have an Nvidia hardware RAID 1 array and need to reformat. But I don't have any spare storage media, so I need to get along with the 2 disks in the array. I figured I'd do as follows: Delete the array, so I have 2 separate but identical disks, A and B, with my files Put disk B aside Reformat disk A, build RAID 1 array with it, install Windows XP Put disk B back in Boot to Windows on disk A, copy my files from disk B to disk A Add disk B to the RAID 1 array, rebuild array And now I'd have a new RAID 1 array, fresh install and all my files intact (the ones I copied). Here are the parts I'm unsure about: Can I build a RAID 1 array using just one disk, then add the other one later? Can Windows on disk A see disk B and allow me to copy my files over?

    Read the article

  • Intraforest user account merge with Active Directory

    - by Neobyte
    I have a scenario where there is a root domain (RD) and two child domains (CD1 and CD2). Users have accounts on both CD1 and CD2, with identical samAccountNames, names etc, and various applications either use the CD1 or CD2 account for authentication to resources. I need to collapse CD2 into CD1, so I want to merge the accounts together. However ADMT does not allow me this option (merge options are greyed out), I think because it does not support intraforest merge of accounts (although it does not explicitly state this anywhere in the documentation). My question is - what is the easiest way for me to merge these accounts? Ultimately all I really need (I think) is for the SID of CD2\user1 to be added to the SIDHistory of CD1\user1 - is there a tool that supports this? Computer accounts and profiles are not a concern for this scenario. Group migration is unlikely to be an issue either - CD2\user1 is usually granted resource access through membership of a group on CD1.

    Read the article

  • Alt-Escape has different effect on different Win-XP machines

    - by Ram Rachum
    This is really weird. On my desktop computer, I often use Alt-Escape to send the active window to the background. This is really useful for window management. However, when I try pressing Alt-Escape on my new laptop, it does something similar but not identical: It sends the active window back, but not all the way to the background. i.e., instead of giving it the lowest index number, it just decrements its index number, probably by 1. Both computers have the same Windows XP Professional. Why is this? And how can I make my laptop computer send the active window to the background instead?

    Read the article

  • why is iometer performance slower after first run?

    - by Dan
    I'm doing some benchmarking with IOMeter, and I'm seeing a consistent and susbtantial drop-off in performance after running the first test. These drop-offs are about the same on the three systems I've tested on, which makes me think it's a configuration setting, or just a fact of life about using IOMeter. For example, one system (local RAID 10) went from 388 I/Os per second the first run, to about 211 I/Os per second on every run after that. Everything else about the test was identical, and I also bounced the machine in between runs. So, is this expected behavior, or am I missing something?

    Read the article

  • configure Heartbeat on Centos Linux - error message

    - by Elad Dotan
    I installed Heartbeat on my Centos Linux and it seems to partially work..but I'm trying to monitor a service with no success. only when I reboot the main server the backup server takes over. in the logs I get : heartbeat[30476]: 2012/03/20_18:51:57 WARN: string2msg_ll: node [node1] failed authentication heartbeat[30476]: 2012/03/20_18:51:58 WARN: string2msg_ll: node [node02] failed authentication the authkeys is identical (copied from one to another). this is my ha.cf: logfile /var/log/ha-log logfacility local0 keepalive 2 deadtime 30 initdead 120 bcast eth0 udpport 694 auto_failback on node server01.com node server02.com haresources : server01.com 38.108.117.3 aim chat any idea how to fix the problem so if a service stops the other server take over Thanks! E.

    Read the article

  • How can I get vim to set an ACL on its swap files?

    - by thsutton
    I use vim on an OS X Snow Leopard Server machine. A number of the directories I work in have ACLs (so that various groups of users can access them over AFP) that are inherited. For some reason, when I'm working in one of these directories, vim cannot read it's own swap files. It can create them fine but can't read them which, for some reason, makes it display the "swap file already exists" message (and no, the swap file does not already exist). vim -r lists the newly created swap file as "[cannot be read]". The owner and group are correct and the permissions are 0600, and the ACLs on the swap file and the file I'm editing are identical (as disclosed by ls -le and compared with diff). groups returns the same thing whether invoked from my login shell or via :! in vim. Has anyone encountered (and hopefully resolved) a problem like this before?

    Read the article

  • mod_access for lighttpd causes a 403 error for all POST requests

    - by Sam
    I have found on my debian server that running the lighttpd module mod_access is causing the server to response with a 403 to all POST requests. It's very odd as I have two servers, one is running as I'd expect and the other keeps returning these 403's. They are running identical configs for lighttpd and php. My lighttpd.conf is: https://gist.github.com/4269500 There is also one other custom conf: https://gist.github.com/4269508 I've opened up the servers for requests until I get this fixed, the server that works is http://mercury.isitup.org/ and the one that fails is http://venus.isitup.org/. After working out that disabling mod_access resolves the problem I greped all my lighttpd configs for uses of it (docs). Disabling each line I found didn't help, leading me to think this is perhaps some default behaviour (or bug?)... Has anyone come across this before or know what configuration value I've got wrong? Versions Debian: Debian GNU/Linux 6.0.6 (squeeze) Lighttpd: lighttpd/1.4.28 (ssl) PHP: PHP 5.3.19-1~dotdeb.0 with Suhosin-Patch (cli)

    Read the article

  • Mixing RAM brands

    - by Andy
    Hey guys, I'm scheduled to install a server at a colo facility tomorrow, but I've hit a snag. ESXi doesn't like the fact that I have 12GB ram in my server (8gb on one side of CPU, 4gb on the other). I currently have 3xCrucial 4gb ECC DDR3 1333mhz sticks (http://www.amazon.com/gp/product/B001TGT7EE/ref=oss_product) and I'm planning on stopping by Fry's on the way to the colo facility so I can pick up another stick so that ESXi will install. Problem is, Fry's doesn't stock Crucial RAM. They do have a stick that I think is identical to my crucial one: http://www.frys.com/product/5915524?site=sr:SEARCH:MAIN_RSLT_PG Will this be dangerous to use in my server? The specs on the RAM are the same as far as I know (ecc registered, 1333mhz, ddr3, 4gb).

    Read the article

  • Distributed, Parallel, Fault-tolerant File System

    - by Eddified
    There are so many choices that it's hard to know where to start. My requirements are these: Runs on Linux Most of the files will be between 5-9 MB in size. There will also be a significant number of small-ish jpgs (100px x 100px). All of the files need to be available over http. Redundancy -- ideally it would provide the space efficiency similar to RAID 5 of 75% (in RAID 5 this would be calculated thus: with 4 identical disks, 25% of the space is used for parity = 75% efficent) Must support several petabytes of data scalable runs on commodity hardware In addition, I look for these qualities, though they are not "requirements": Stable, mature file system Lots of momentum and support etc I would like some input as to which file system works best for the given requirements. Some people at my organization are leaning towards MogileFS, but I'm not convinced of the stability and momentum of that project. GlusterFS and Lustre, based on my limited research, appear to be better supported... Thoughts?

    Read the article

  • How to get HTTP preseed to work correctly on Ubuntu 10.04 LTS (Lucid)?

    - by netvope
    Installation media: ubuntu-10.04-desktop-i386.iso I tried a lot of different boot parameters, but either the installer ignored the preseed configuration, or it boot itself directly as LiveCD. An example of the boot parameters I've tried: auto url=http://mydomain.com/path/preseed.cfg boot=casper only-ubiquity initrd=/casper/initrd.lz quiet splash -- If I remove only-ubiquity, it boots as a LiveCD. If I remove boot=casper, it won't boot. If I add vga=normal locale=en_US console-setup/layoutcode=us console-setup/ask_detect=false interface=auto, it still can't do automatic install. If I remove auto, it's the same. What is the correct boot parameters for launching such an installation? From the apache log of the server hosting preseed.cfg, I see that the installer has no problems fetching the preseed file. My preseed file is almost identical to the one at https://help.ubuntu.com/10.04/installation-guide/example-preseed.txt. Moreover, I have run debconf-set-selections -c preseed.cfg to ensure that the preseed file is correct.

    Read the article

  • Keepalived takes several minutes to recover in a particular situation

    - by NathanE
    I've setup Keepalived for a master-slave style virtual IP and it seems to work well. Both are hosted in almost identical VMs. If I "pause" the VM that is running the Master. The Slave will take over, as expected, almost instantly. However if I then "unpause" the VM that runs the Master. The virtual IP will stop responding the pings. And it takes a good 4 or 5 minutes for it to start pinging again. It seems to be getting desynchronised due to the nature of the way I'm testing it (by pausing/unpausing the VMs). I admit that pausing and unpausing VMs is a slightly dodgy way to test this. But it has raised a concern for me that there could be other scenarios that cause the same undesirable behaviour. Is this expected / by design? Is there anything I can do to the config to improve it? Thanks.

    Read the article

  • Are periodic full backups really necessary on an incremental backup setup?

    - by user2229980
    I intend to use an old computer I have as a remote backup server for myself and a few other people. We are all geographically separated, and the plan is to do incremental daily backups using rsync and ssh. My original idea was to make one initial full backup then never again have to deal with the overhead of doing it, and from that moment on only copy the files changed since the last backup. I've been told that this could be bad, but I fail to understand why. Since each snapshot is comprised of hard links to the unchanged files plus the original changed ones, isn't it going to be identical to a new full backup? Why would I want to make another full backup?

    Read the article

  • chmod -R 777 /. - RHEL 5.5

    - by user1263746
    A shell script testing went bad and it issued chmod -R 777 /. to the system, instead of chmod -R 777 ./ and as expected it wiped the critical meta data. We have turned off the system and it will not function properly the next time it is turned on. I am told that rpm --setperms -a rpm --setugids -a should atleast fix the permission of the packages maintained by rpm. Is it worth doing? And is there any script available which will copy the permission from an identical system? To atleast get the box working. The Box is running RHEL5.5 Thanks!

    Read the article

  • How to set path of Virtual PC hard disk differencing parent

    - by Barry Kelly
    I have an old Windows XP Mode vhd backed up from my previous system, but I'm having difficulty getting it running on the new system. The vhd is a differencing disk, and its parent is the standard Windows XP Mode base; I still have the old parent, and have verified it is binary identical to the XP Mode base in my new installation of XP Mode. But in the new system, the path to the differencing disk parent is different than the old. When I open up the settings for the .vmcx for my old XP mode, and select "Hard Disk 1", the "Virtual hard disk file" is set correctly, but the "Parent Disk:" field is pointing at the wrong path, and I can't see any way to edit it. Does anyone know how?

    Read the article

< Previous Page | 21 22 23 24 25 26 27 28 29 30 31 32  | Next Page >