Search Results

Search found 6917 results on 277 pages for 'upgrade osso to oam 11g'.

Page 193/277 | < Previous Page | 189 190 191 192 193 194 195 196 197 198 199 200  | Next Page >

  • cannot connect to MS FTP 7.5 on Windows 2008 on Amazon EC2 instance

    - by minerj
    I have just installed the MS FTP 7.5 upgrade on my Windows 2008 Server (Service Pack 2) running on an Amazon EC2 instance. In the FTP Firewall Support settings for the server in IIS Manager I have set up the passive port range 45001 - 45005 and also set the External Firewall IP address to match the assigned Amazon Elastic IP address. Using the AWS Console I changed the Security Group for the server to allow access to the server through ports 21 and 45001 through 45005. Using an FTP client (either the command line FTP client or Windows Explorer) on the Amazon server I can connect to the FTP server but I cannot connect with an external FTP client. When I checked to see which ports were open on the server using Shields Up it shows that port 21 is open but ports 45001 to 45005 are closed. I assume I'm missing something. Any help greatly appreciated.

    Read the article

  • SQL 2008 publisher -> SQL 2000 subscriber: Is a pull subscription possible for merge replication?

    - by Brian Dunzweiler
    I am trying to synchronize a SQL 2000 SP4 subscriber to a SQL 2008 publisher via a merge pull subscription. When the subscriber tries to run the merge agent, it fails the following error: The process could not connect to Distributor 'OH05DBS002\SAM_SSG_2008'. SQL Server does not exist or access denied. Has anyone had success with this setup? I was able to create and synchronize a push subscription so I know that communication works between the two, at least from 2008-2000. The lack of communication from 2000-2008 also affects the ability to create a linked server on the SQL 2000 subscriber. One other tidbit - I did install the SQL 2008 native client on the the 2000 box but it didn't help either. Before anyone asks, I can't upgrade the subscriber as it still needs to support replication between MS Access 2003. Yeah, I know. :) TIA, Brian

    Read the article

  • Puppet yum repo - Pull down 2.7.x vs 3.0.x

    - by Mike Purcell
    So a few weeks ago I started on the path to using puppet to automate all the configs/services. At the time I was using the EPEL repo, which installed version 2.6.x. After some reading I was trying to gain access to the flatten method available via the puppet stdlib, and thought it was available by default in the newer 2.7.x version. So I added a puppet repo with the following settings: [puppetlabs] name=Puppet Labs Packages baseurl=http://yum.puppetlabs.com/el/$releasever/products/$basearch/ enabled=1 gpgcheck=1 gpgkey=http://yum.puppetlabs.com/RPM-GPG-KEY-puppetlabs The problem with this, is it installed v3.0.x instead of 2.7.x. And apparently 3.0.x is a major upgrade which was released only a few weeks ago. Obviously I would prefer to use the 2.7.x for the next few months while PuppetLabs fix any defects which will inevitably arise after a major version. So my question is, what setting can I add to the puppet repo config to pull down only the 2.7.x branch and not the 3.0.x branch?

    Read the article

  • How to install plug-in for Google Chrome

    - by Jeff
    Recently Google Chrome browser has been prompying me to install a plug-in everytime visit a web page. I always say Yes, install pulg-in, but that seems to have no effect. I tried following the "Trouble installing plug-ins" on the Chrome toolbar, but that seems to say Windows Media Player is the problem, but again, all my attempts at installing don't have any effect. As far as I know, I have not chnage anything, but Skype did recently upgrade itself. This is Windows 7 Professional 64-bit, and Chrome says it is up-to-date. I'm going to run a malware checker next, just in case - Thanks!

    Read the article

  • How do I create an ISO image from a directory structure on CentOS?

    - by tom smith
    I'm trying to figure out the exact mkisofs cmd to create the ISO with the following directory and file structure. I've tried different commands, but when I mount the ISO that is created the directory tree has not been reproduced. The initial directory tree is: master.iso:: mount -o loop /apps/vmware/master.iso /mnt/vmtest ls /mnt/vmtest isolinux ks.cfg upgra32 upgra64 upgrade.sh ls /mnt/vmtest/isolinux boot.cat initrd.img isolinux.bin isolinux.cfg vmlinuz I've used different variations of the following mkisofs command without success: mkisofs -o '/foo/test.iso' -b 'isolinux.bin' -c 'boot.cat' -no-emul-boot -boot-load-size 4 -boot-info-table 'isolinux' How do I make an ISO that captures a directory's exact structure?

    Read the article

  • Packaging python software with custom dependencies

    - by viraptor
    Hi, I'm looking for a good way to package a Python application that is going to be deployed on a Debian server. The application itself depends on some modules which are not included in base Debian repository, although they might be in the future. This creates some problems... I depend on some patches to those modules. If the original module gets installed one day, the application will break. However if I install everything I need in a virtualenv just for that application, I lose the ability to upgrade Python itself (in case of security updates). The third option would be to rename my fork of the upstream module and just treat it as a completely separate one. But that would mean changing the code (not much work, but it wouldn't be that clean / universal anymore). Are there any other options that I missed? Are there any pros / cons I didn't see in the solutions above?

    Read the article

  • Trigger ZFS dedup one-off scan/rededup

    - by Jake Wharton
    I have a ZFS filesystems which has been running for some time and I recently had the opportunity to upgrade it (finally!) to the latest ZSF version. Our data doesn't scream dedup but I firmly believe based on small tests that we could gain anywhere from 5-10% of our space back for free by utilizing it. I have enabled dedup on the filesystem and new files are slowly being dedupified but the majority (95%+) of our data already exists on the filesystem. Short of moving the data off-pool and then recopying it back, is there any way to trigger a dedup scan of existing data? It doesn't have to be asynchronous or live. (And FYI there isn't enough room on the pool to copy the entire filesystem to another and then just switch the mounts.)

    Read the article

  • Get sudoers information for all users on a server

    - by Jon Kruger
    I want to collect information on what sudoers actions a user can perform on a server. From what I've read, you can do this with the command sudo -l -U username. However, one server that I have has a slightly older version of sudo (1.6.7p5) and the -U option doesn't seem to exist (I don't own the server so I can't just upgrade to a newer version of sudo). Has anyone ever had to collect sudoers information for all users on a server? How would you recommend doing it?

    Read the article

  • Ways for managing the installation and configuration of various software applications and settings i

    - by EmpireJones
    What are some ways for managing the installation and configuration of various software applications and settings in group of linux development and server computers? Is a set of basic scripts a good means of attack? I was thinking about just having a ton of scripts, such as: setup_dev_env [install|uninstall|reinstall] setup_nfs [...] setup_nfs_share [...] setup_http [...] setup_memcache_node [...] Is there any better method? It would be nice to be able to "upgrade" an installation script too, for example, to change common development settings.

    Read the article

  • Apache availability with the two front-ends on diferent locations. Is it possible?

    - by marc.riera
    Hello, I have to locations (office and service providers). One DNS(bind) serving our domain as authoritative, and a service provider webserver with our corporate web on a private server. So.. Now we are planing to upgrade our server on the ISP to a new one, and I would like to use this situation to improve our service. Is it possible to mount a high availability apache/mysql/php within to different locations? I will install a bind slave on the same new server, so I hope it will make things easier, but I need some hints and tips on how to ride it. THanks.

    Read the article

  • How to set up raid monitoring on a Adaptec AAC-RAID on a Dell?

    - by pjz
    I've got Debian on a Dell with an Adaptec RAID: AAC0: kernel 4.1-0[7417] AAC0: monitor 4.1-0[7417] AAC0: bios 4.1-0[7417] AAC0: serial 4edf09 scsi2 : aacraid Vendor: CERC Model: DATA 1 Rev: V1.0 Type: Direct-Access ANSI SCSI revision: 02 Vendor: CERC Model: Data 2 Rev: V1.0 Type: Direct-Access ANSI SCSI revision: 02 I've got it set up using the fine afacli tool (ick). Is there a way to get it to email me when there's an error? Do I need to hand-roll a script for this? what's the right version of the afacli tool to use? v4.1 gets SIGABRT and exits after I ask it to 'container list' or 'enclosure list' v2.8's enclosure commands don't work either Do I need to upgrade firmware? to what? from where? how?

    Read the article

  • Cold Fusion 8 and 9 on IIS 6

    - by David Mesh
    I have a client, who against my better judgement has insisted on doing the following. They have a single IIS 6 on Win 2003 Server Enterprise. It currently runs ColdFusion 8. They want me to install ColdFusion 9 on the server without changing any of the existing sites so that they can develop in CF 9 and upgrade other sites in the future. Yes, I have begged them to use a Dev server, or run on apache on the same box. Can this even be done? Many thanks in advance! DM

    Read the article

  • needs updated glibc package version 3.4.15 or later for RHEL6

    - by Tejas
    I want to upgrade my current running applications to latest version. But due to some package issue i am unable to install them. I get common error in that: /usr/lib64/libstdc++.so.6: version 'GLIBCXX_3.4.15' not found. When i tried to update glibc package i get following output: [root@agastya ~]# yum install glibc Loaded plugins: refresh-packagekit, rhnplugin epel/metalink | 3.8 kB 00:00 epel | 4.3 kB 00:00 epel/primary_db | 5.0 MB 01:33 epel-testing/metalink | 3.8 kB 00:00 epel-testing | 4.3 kB 00:00 epel-testing/primary_db | 295 kB 00:03 rhel-x86_64-server-6 | 1.8 kB 00:00 rhel-x86_64-server-6/primary | 11 MB 02:02 rhel-x86_64-server-6 8816/8816 Setting up Install Process Package glibc-2.12-1.80.el6_3.6.x86_64 already installed and latest version Nothing to do [root@agastya ~]# Should i need to add some more repositories? If yes, how?

    Read the article

  • SSL to SSL Redirects in IIS - Possible?

    - by Eric
    We have a situation where we would like to redirect https://service1.domain.com to https://service2.domain.com. I know this is very simple with http endpoints, but I'm not too sure about https. We have some legacy windows application web service clients that will not be updating their software version soon, and we cannot update their web references to https://service2.domain.com. Is there any way to leave these web service clients pointing to https://service1.domain.com, but have their requests forwarded to (and responded to by) https://service2.comain.com? The old server is running IIS 6.0. The new server is running IIS 7.0. We could probably upgrade it to 7.5 if needed, but I'm not certain. We could also probably make a seamless transition of the old web service to a new server using public DNS, but we cannot change the DNS name of "service1.domain.com." Thanks ServerFault!

    Read the article

  • Debian no lib32asound2-plugins=1.0.25

    - by erai
    I compiled git wine 1.5.6 and the sound is not working correctly. winecfg gives a message: ALSA lib ../../pulse/pcm_pulse.c:995:(_snd_pcm_pulse_open) Unknown field handle_underrun err:winediag:AUDDRV_GetAudioEndpoint PulseAudio "default" -22 without handle_underrun. Audio may hang. Please upgrade to alsa_plugins >= 1.0.24 err:ole:CoInitializeEx Attempt to change threading model of this apartment from multi-threaded to apartment threaded But on wheezy there is no package lib32asound2-plugins 1.0.25 So, how do I build and package alsa-plugins to install to lib32? Or how do I use automake's configure to compile a 32bit lib on a 64 bit system. All I Need now is the libavcodec-dev for 32bit, but I don't understand. What is the 32bit libavcodec-dev package for amd64 debian?

    Read the article

  • suggestions for migrating a windows 7 install to a new 4K sector disk

    - by myCubeIsMyCell
    Hi, I'm looking to upgrade the disk on a windows 7 box to a new larger drive. In the past for such migrations I'd just hook both drives up and use a linux boot disk and use dd to copy from one disk to the next... boot up the new drive & expand the partition. The drive I just purchased however is a western digital using 4k sectors... not sure if there'd be any complications using my old method moving from a 512b sector drive. Current plan is to try the migration by doing a win7 system image backup to an external drive... then restore the image to the new drive via system restore boot disk. Any suggestions or recommendations on how to best complete this migration would be greatly appreciated. Thanks!

    Read the article

  • Moving mails from SBS 2008 to SBS 2011

    - by starchx
    we are going to upgrade server from SBS 2008 to a new hardware server with SBS 2011. Due to non-experience in Server Migration (eg, answer file), we have decided to choose Clean Install. We have 12 users so that's not too bad. The most challenge problem is to move over all the emails for each user, some user have 5GB inbox. The one method I know is to login to each user account and export the emails to PST file using Outlook 2007. Then connect to new SBS 2011 and import the PST into new inbox. Do we have any other better/quicker ways? Thanks

    Read the article

  • Integrated Graphics and Audio as Media Center?

    - by Will
    I'm considering setting up a PC as a Media Center. Mainly to watch movies (ideally HD quality) and listening to music, but also to perform tasks like e-mail, web browsing, ... I quite like the looks and the price of this barebone: http://www.asus.de/Barebone_PC/S_Series_7L/S2P8H61E However it comes with integrated graphics and audio and only has one free PCI-Express slot. Which would mean in the worst case, where both integrated graphics and audio turn out to be insufficient, I could only upgrade one. So is integrated graphics and audio sufficient for a media center solution? Cheers, Will

    Read the article

  • use yum behind a proxy on CentOS 5.6

    - by ewok
    I have a CentOS VM running in Virtualbox on a Windows 7 host (all 64-bit). I have configured the proxy in the System menu and have added the proxy=http://my.proxy.com:12345 line to my /etc/yum.conf file. I am able to connect to the internet from a browser, and I can update packages form the update manager and install them from the package manager. However, I cannot use yum for anything. When I attempt to use yum (install, update, upgrade, or anything else), It attempts to connect to the mirrors, does nothing, then after a minute or so, quits. Does anyone know how to configure yum to work behind a proxy?

    Read the article

  • Setting up a Time Capsule with port forwarding

    - by Kaji
    Our old AirPort Extreme station hit EOL, so we decided to upgrade it to a Time Capsule. Along the way, we're trying to also set it up with a separate guest network and port forwarding/NAT, however we're having trouble setting it up so that the time capsule is handling the DHCP leases instead of the router. We've got DSL through Verizon through a Westell modem/router to the Time Capsule. Done the RTFM thing, and we haven't been able to get it to work. Can anyone explain how to get things set up properly for this configuration?

    Read the article

  • Setting up a Time Capsule with port forwarding

    - by Kaji
    Our old AirPort Extreme station hit EOL, so we decided to upgrade it to a Time Capsule. Along the way, we're trying to also set it up with a separate guest network and port forwarding/NAT, however we're having trouble setting it up so that the time capsule is handling the DHCP leases instead of the router. We've got DSL through Verizon through a Westell modem/router to the Time Capsule. Done the RTFM thing, and we haven't been able to get it to work. Can anyone explain how to get things set up properly for this configuration?

    Read the article

  • Fresh install of 64bit Ubuntu needs flash but adobe doesn't have a version for me.

    - by DJTripleThreat
    I need flash to watch YT videos. YT said "You need to upgrade your Adobe Flash Player to watch this video." with a link to download flash. I'm running 10.04 so I see possible choices for myself: 1) a .deb file for Ubuntu 8.04+ or APT (whats this??) for Ubuntu 9.04. I downloaded the deb file and when I opened it in the installer it said that I have the wrong architecture. Anyone have an idea about how to work around this?

    Read the article

  • pip install very slow through virtual box

    - by AJP
    pip install --exists-action=w -r requirements.txt is very very slow through virtual box. Any suggests of how to diagnose and fix? Would seeing the VagrantFile be useful? VirtualBox 4.2.12 (can't upgrade to .14 as it doesn't work.) Vagrant 1.0.7 Host machine: ProductName: Mac OS X ProductVersion: 10.7.5 BuildVersion: 11G63b VagrantFile contains: Vagrant::Config.run do |config| config.vm.box = "precise64" config.vm.customize ["modifyvm", :id, "--memory", 2048] config.vm.box_url = "http://files.vagrantup.com/precise64.box" config.vm.network :hostonly, "33.33.33.21" config.vm.forward_port 5000, 5000 config.vm.forward_port 5555, 5555 config.vm.share_folder "v-root", "/vagrant", "./" Vagrant::Config.run do |config| config.vm.provision :shell, :inline => "VENV=/usr/local/venv bash /vagrant/setup_env.sh" end end Normal download speed is only about 5 times slower at 0.8 Mb per second versus 4 MB per second (as judged by curling a 50 Mb file from S3). But pip install is taking about 20 times longer from Mac (i.e. about 40 minutes) versus 2.

    Read the article

  • How to find out if the screen is WLED or RGBLED on Dell Studio XPS 16?

    - by Abhijeet
    I just ordered the Dell Studio XPS 16 with i7 and 16" RGBLED screen. This upgrade from WLED LCD to RGBLED LCD charged me more $$. But now when I view the order online, it lists this part as "320-8335 Premium FHD WLED Display, Obsidian Black, 2.0 MP Webcam". When called Dell, the rep says this part is RGBLED screen and the "premium" is for RGBLED. I want to make sure they are shipping me the RGBLED screen and not the WLED. Is there any way to verify this after I receive (either from Device Manager or BIOS or somewhere in system settings)? Also, is there any published specifications / criteria that we can run (some third party software) on this monitor which can tell if this is WLED or RGBLED?

    Read the article

  • What is the peak theoretical WiFi G user density? [closed]

    - by Bigbio2002
    I've seen a few WiFi capacity planning questions, and this one is related, but hopefully different enough not to be closed. Also, this is related specifically to 802.11g, but a similar question could be made for N. In order to squeeze more WiFi users into a space, the transmit power on the APs need to be reduced and the APs squeezed closer together. My question is, how far can you practically take this before the network becomes unusable? There will come a point where the transmit power is so weak that nobody will actually be able to pick up a connection, or be constantly roaming to/from APs spaced a few feet apart as they walk around. There are also only 3 available channels to use as well, which is a factor to consider. After determining the peak AP density, then multiply by users-per-AP, which should be easier to find out. After factoring all of this in and running some back-of-the-envelope calculations, I'd like to be able to get a figure of "XX users per 10ft^2" or something. This can be considered the physical limit of WiFi, and will keep people from asking about getting 3,000 people in a ballroom conference on WiFi. Can anyone with WiFi experience chime in, or better yet, provide some calculations for a more accurate figure? Assumptions: Let's assume an ideal environment with no reflection (think of a big, square, open room, with the APs spaced out on a plane), APs are placed on the ceiling so humans won't absorb the waves, and the only interference are from the APs themselves and the devices. As for what devices specifically, that's irrelevant for the first point of the question (AP density, so only channel and transmit power should matter). User experience: Wikipedia states that Wireless G has about 22Mbps maximum effective throughput, or about 2.75MB/s. For the purpose of this question, anything below 100KB/s per user can be deemed to be a poor user experience. As for roaming, I'll assume the user is standing in the same place, so hopefully that will be a non-issue.

    Read the article

< Previous Page | 189 190 191 192 193 194 195 196 197 198 199 200  | Next Page >