Search Results

Search found 11861 results on 475 pages for 'methods rec'.

Page 237/475 | < Previous Page | 233 234 235 236 237 238 239 240 241 242 243 244  | Next Page >

  • SYN flooding still a threat to servers?

    - by Rob
    Well recently I've been reading about different Denial of Service methods. One method that kind of stuck out was SYN flooding. I'm a member of some not-so-nice forums, and someone was selling a python script that would DoS a server using SYN packets with a spoofed IP address. However, if you sent a SYN packet to a server, with a spoofed IP address, the target server would return the SYN/ACK packet to the host that was spoofed. In which case, wouldn't the spoofed host return an RST packet, thus negating the 75 second long-wait, and ultimately failing in its attempt to DoS the server?

    Read the article

  • Better to develop Ruby project on server or buy a faster desktop computer? B/c laptop too slow.

    - by user33184
    I have a Linux laptop (Vostro V13) running a Celeron M chip. This is a fine laptop, but running unit tests especially for Rails applications is slow. I want a faster development environment but I don't want to spend too much money. So the choice I have is between $390 for a Linux desktop machine with a Pentium Dual Core Processor E5400 and just paying between $30 and $40 a month to Linode and trying to do development remotely on that server. Can anyone with experience developing server applications using both methods offer any advice?

    Read the article

  • Transferring Microsoft CRM data

    - by notCRMguru
    My boss has asked me to transfer data from the current Microsoft CRM 4.0 server to a new one. I myself haven't used CRM at all. I've done some research and come across various ways to import data from different sources. These methods include using CSV files and Data Maps. This seems very cumbersome and unnecessary since the data is already in a CRM. Would someone please direct to some guides for full/partial data transferral from this current CRM to a new one? Thanks

    Read the article

  • Puppet, Nagios, Munin on cPanel based hosts

    - by WinkyWolly
    I've been managing 20-30~ cPanel based hosts over the past year with Puppet, Nagios and Munin for general monitoring / trending however a lot of the methods I've had to use to deploy / manage things such as configurations a pain. For those of you who aren't familiar with cPanel - it adds a few things to yum exclude such as perl*, ruby* and so forth. This causes issues with me being able to bootstrap monitoring on a new server via Puppet (well via the Package type) due to a bunch of conflicts with installing via Yum. Now I could create a custom RPM for everything and remove certain dependencies from the spec file however I would like to avoid this if possible. Does anyone have any proposed functional ways to manage this sort of environment? Currently I install Puppet, Facter and Munin via RPM's and force install using --nodeps and such (since they're installed, just no the ones Yum wants). Nagios I installed manually from source at this time (likely will create RPM's however I want to tackle this general issue first).

    Read the article

  • SQL Server 2008 R2 100% availability

    - by Mark Henderson
    Is there any way to provide 100% uptime on SQL Server 2008 R2? From my experience, the downtimes for the different replication methods are: Log Shipping: Lots (for DR only) Mirroring w. NLB: ~ 45 seconds Clustering: ~ 5-15 seconds And all of these solutions involve all of the connections being dropped from the source, so if the downtime is too long or the app's gateway doesn't support reconnection in the middle of task, then you're out of luck. The only way I can think to get around this is to abstract the clustering a level (by virtualising and then enabling VMWare FT. Yuck. Good luck getting this to work on a quad-socket, 32-core system anyway.). Is there any other way of providing 100% uptime of SQL Server?

    Read the article

  • Can I rely on S3 to keep my data secure?

    - by Jamie Hale
    I want to back up sensitive personal data to S3 via an rsync-style interface. I'm currently using s3cmd - a great tool - but it doesn't yet support encrypted syncs. This means that while my data is encrypted (via SSL) during transfer, it's stored on their end unencrypted. I want to know if this is a big deal. The S3 FAQ says "Amazon S3 uses proven cryptographic methods to authenticate users... If you would like extra security, there is no restriction on encrypting your data before storing it in Amazon S3." Why would I like extra security? Is there some way my buckets could be opened to prying eyes without my knowing? Or are they just trying to save you when you accidentally change your ACLs and make your buckets world-readable?

    Read the article

  • nginx redirect TLD to TLD with virtual folder (example.com => example.com/test)

    - by Amund
    Im running nginx and in the config file I need to always have the domain example.com redirect to example.com/test. I tried various methods for achieving this but I always got a redirect error. What is the correct way to do this? nginx.conf snippet: server { server_name example.com www.example.com; location / { rewrite ^.+ /test permanent; } } server { listen 80; server_name www.example.com example.com; location / { root /var/www/apps/example/current/public; passenger_enabled on; rails_env production; } } Thanks!

    Read the article

  • datacenter network change control best practices

    - by jpolache
    I have been tasked with compiling a list of possible network equipment changes at a data center. The task includes tagging which changes need change control and which don't. Does anyone know of a "best practices" list that I can start from? The methods for doing change control at this data center are well established. The list would be of specific configuration items that should or should not be included in the change control process, of example; static route entries switch port assignments firewall rule additions/changes etc.

    Read the article

  • .htaccess configuration issue

    - by Hammad Haider
    Hi, i am using two website on one domain like: www.example.com & www.example.com/site2, i want to know that on my site2, in my site2 their are 2 folders name folder1 and folder2 my index.php is in folder2 but the defination of methods defined in folder2 i am including the files through .htaccess but i am unable to get those files which are in folder1 and getting Error-500 and 400 on browser and i am using following lines but they are not working in .htaccess file The line below works fine RedirectMatch ^/$ http://www.example.com.pk/site2/views/ AllowOverride All php_value include_path ".:/home/example/public_html/site2/system" waiting for your quick response. Thanks Regards, Hammad Haider.

    Read the article

  • Cage nuts - which way to insert them?

    - by Ingmar Hupp
    Cage nuts can be inserted (in square holes) either with the flaps horizontal or vertical. I've seen and used both methods, but am still not sure which is best. Is there a recommended way of inserting them? If so, why is it the best way? From my own experience, the horizontal way (on the left) seems slightly better for no other reason than making it easier to insert or pry out the cage nut in a full rack (with the units above and below being populated), for example when replacing cage nuts with a snap-in rail. Edit: Major vendors each seem to have their own preference too: Horizontal: IBM, Cisco, APC Vertical: HP (and Compaq), Sun/Oracle, Dell

    Read the article

  • Open file with a batch file pinned to the taskbar

    - by daniel
    How do you get a file opened by a batch file pinned to the taskbar? In XP you could just drag your batch file onto the quicklaunch section of the taskbar, then any file you drop on that icon would be opened with that batch file. In windows 7 you can get a batch file there using various methods, and you can open a file with a pinned application by shift dragging onto it, but this doesn't work for batch files. related article: How to pin either a Shortcut or a Batch file to the new Windows 7 Taskbar?

    Read the article

  • Apache + LDAP Auth: access to / failed, reason: require directives present and no Authoritative hand

    - by Karolis T.
    Can't solve this one, here's my .htaccess: AuthPAM_Enabled Off AuthType Basic AuthBasicProvider ldap AuthzLDAPAuthoritative on AuthName "MESSAGE" Require ldap-group cn=CHANGED, cn=CHANGED AuthLDAPURL "ldap://localhost/dc=CHANGED,dc=CHANGED?uid?sub?(objectClass=posixAccount)" AuthLDAPBindDN CHANGED AuthLDAPBindPassword CHANGED AuthLDAPGroupAttribute memberUid AuthLDAPURL is correct, BindDN and BindPassword are correct also (verified with ldapvi -D ..). Apache version: Apache/2.2.9 (Debian) The error message seems cryptic to me, I have AuthzLDAPAuthoritative on so where's the problem. EDIT: LDAP modules are loaded, the problem is not with them being missing. # ls /etc/apache2/mods-enabled/*ldap* /etc/apache2/mods-enabled/authnz_ldap.load /etc/apache2/mods-enabled/ldap.load EDIT2: Solved it by changing funky Require ldap-group cn=CHANGED, cn=CHANGED line with Require valid-user Since AuthzLDAPAuthoritative is on, no other auth methods will be used and valid-user requirement will auth via LDAP. (right? :/)

    Read the article

  • SSH broken after homedir permissions and hostname change on EC2-hosted Ubuntu

    - by dimadima
    I changed my instance's hostname using the hostname utility and then set it in /etc/hostname so that the new name survives reboot. My main motivation was for differentiating between instances at the prompt using the \h format in PS1. EDIT I also changed permissions on my home directory. I made my home directory group writeable. Now I can no longer SSH into the machine. The short of it is the error Permission denied (publickey). Running ssh -v, the more verbose output is: debug1: Authentications that can continue: publickey debug1: Next authentication method: publickey debug1: Offering RSA public key: /Users/dmitry/.ssh/id_rsa debug1: Authentications that can continue: publickey debug1: Trying private key: /Users/dmitry/.ssh/ec2key.pem debug1: read PEM private key done: type RSA debug1: Authentications that can continue: publickey debug1: No more authentication methods to try. Permission denied (publickey). Should I have done something after changing the hostname? Now I can't get into the instance! :(

    Read the article

  • Unable to open Synaptic manager, unable to install packages, what can I do?

    - by Omkant
    I have installed the sun java6-jdk package but it's not installed completely. After that it's giving an error when I try to install a new package. In terminal when I write "sudo apt-get install [any package name]" I get an error like this: E: could not get lock /var/lib/dpkg/lock -open (11 Resource temporable ) E: Unable to lock the administration directory (/var/lib/dpkg). Also Synaptic package manager is not opening and none of the packages are downloading through any of the methods I know. Please help!

    Read the article

  • Oracle Advanced Security Options is Blank

    - by mak4pi
    I just installed Oracle DB 10gR2 with Oracle Advanced Security, but cannot see the algorithms. [user@db-1] adapters Installed Oracle Net transport protocols are: IPC BEQ TCP/IP SSL RAW Installed Oracle Net naming methods are: Local Naming (tnsnames.ora) Oracle Directory Naming Oracle Host Naming Oracle Names Server Naming Installed Oracle Advanced Security options are: Where are all the algorithms for Oracle Advanced Security options please? I checked the $ORACLE_HOME/bin/adapters file and it's looking for naea256i, naemd5i, etc. in the naetab.so file, but none of these are listed in the naetab.so file. What's wrong with the naetab.so file? Thanks.

    Read the article

  • Free, Linux-based rescue CD for Windows machines

    - by Adam Matan
    Hi, Too often, I'm being called to help a friend who screwed a Windows machine by some creative methods. Th usual remedy is backing up the hard drive contents and reinstalling. Right now, this is done by removing the defected hard drive to my machine. I figured out that using a rescue disk running some version of Linux might ease the process. I'm looking for: NTFS access Partition tools Large variety of drivers (Network, Hard drives, etc.) GUI and some rescue wizards a great plus. Any ideas? Adam

    Read the article

  • Juniper SRX1400 VPN

    - by ank
    I have been trying to set up a client VPN on a Juniper SRX1400 without much success. All documentation I found from Juniper and elsewhere does a lot of other (difficult and wonderful) things other than the simple things I want to do. We already have a VPN running on CISCO PIX hardware which we need to put to rest and we pretty much like to replicate the functionality, which is: 1) client makes a request to the outside interface of the SRX1400, 2) client gets authenticated, 3) if successful, then client is assigned all the usual DHCP stuff and becomes part of the network. What privileges this DHCP assigned network has, should be configurable of course with the usual routing/filtering methods. Am I asking for too much from the SRX for this kind of thing? Should I ditch the SRX for VPN and revert to an OpenVPN solution that I had working some time back also and was amazed at the ease of configuration, functionality and features?

    Read the article

  • How to discern video details from Linux Live CD

    - by Chris Lieb
    I was having trouble getting X.Org to work with the video card in a recently aquired laptop, so I decided to try to probe around the configuration of a Linux Live CD to figure out how to set up my kernel. I chose the Sabayon 5.0 Live DVD because it was the only one that I could find that was the same or newer than the 2.6.32 kernel I have installed on the laptop right now. Of course, the DVD booted right into GNOME without issue, unlike the hangs and crashes that I had been getting. So, I want to figure out what kernel driver this Live DVD is currently using, but I don't know how to. hwinfo doesn't shed any light on the matter, and lspci -vnn doesn't list a driver for my video device, though both methods correctly identify that I have an Intel 855GM video processor. How can I find out this information?

    Read the article

  • SQL Server Management Studio Connect to Server List Editing

    - by Paul Farry
    I'm using SQLServer Management Studio (2005) and I have a fairly lengthy list of servers in there, and I'd like to get rid of some of them that are no longer in use, without having to set them all up again. I know that the C:\Users\*\AppData\Roaming\Microsoft\Microsoft SQL Server\90\Tools\Shell\mru.dat can be deleted and this will remove ALL the entries, but is there anyway to just delete some of them? (Coding info) I looked at the file and it is a serialised blob from the Microsoft.SqlServer.Express.ConnectionDlg.dll (Class Personalization) in the Appplication directory, but all the methods are private. So I can't just create an instance of this and then call Remove on the entries. Update I have written an Article on CodeProject explaining How this can be achieved. http://www.codeproject.com/KB/vb/AlterSQL2005MRU.aspx

    Read the article

  • Checking for orphaned snapshots - ESXi5

    - by Tim Alexander
    So we had some issues with our passive mail node over the weekend doing vmtools updates and to resolve a problem we had to revert to a snapshot and then reseed all the databases across. All in all everything seemed fine, the server works and CCR copy status is running fine. I used the "Delete All" option this morning to remove the snapshot and the process according to vCenter has completed with no errors and no "Needs Consolidation" flag. This all seems fine until I check the Datastore that holds the VM on our SAN and I can clearly see snapshots that are pretty big [see attached image]. These do not seem to be changing size and the data modified is around the time the works were started for the vmtools update. Does this possibly mean that at some stage, possibly during reversion or hard resetting of the VM, that they have become orphaned? Are there any methods to check orphaned status of snapshots? We are running ESXi5.0 Update 1 with storage provide by an EMC SAN. Enterprise plus is the license level.

    Read the article

  • Resolving real IP address out of a dynamic dns address

    - by stavnir
    I recently opened a dynamic dns user (at no-ip for that matter..) for my own personal needs and especially for ssh-ing my computer whenever I need to, without knowing it's static IP. My questions are: Am I misusing the concept of dynamic dns? Are there more appropriate methods to do what I want to do? If not, how do I resolve my router's real ip address? Firefox somehow manages to do so, nslookup and other similar commands only resolve the ip of the ddns server (e.g. no-ip.org). Trying to figure this mystery with wireshark failed miserably ;)

    Read the article

  • Web and email host migration - Limitations and suggestions to make the process as easy as possible.

    - by Jack Hickerson
    I developed a website for a friend of mine to replace his current 'all inclusive' provider (website creation, updating, web hosting, email hosting). I've already paid for a hosting service which currently houses the website which I have created. I need to cancel the previous service provider to get the domain migrated to the new host, however I will still need to transfer or recreate all of the email addresses that everyone in his company had previously. Is there an easy way migrate email accounts (still linked to the same domain) while migrating to a different host? Will any methods allow all users to retain their archived emails and folder structures? What is the process to do so. Because the current provider is a rather large website development and hosting company, I will have limited access to the data they have stored. As you can probably tell, my knowledge in this area is very limited - any/all suggestions you may have would be greatly appreciated. Thanks in advance. -Jack

    Read the article

  • EC2-like virtualisation platform with non-RDP access

    - by code'
    I'm looking into setting up a few small VMs. The trouble is the software I intend to install on them (Cisco VPN Client) blocks out networking (other than to the target VPN destination...) with no workaround. This means that Remote Desktop or other methods of connecting to VMs that go via the Internet (e.g. GoToMyPC, LogMeIn) are a non-starter. What I'm really looking for is an EC2-like platform but which gives direct access to the VM through (for example) Hyper-V Manager. Sadly the only way they all seem to offer to remote control the VMs is direct access via Remote Desktop, whereas I need to be one layer above that (if that makes sense). A viable alternative would be to run virtualisation software within a Windows EC2 instance; obviously hardware virtualisation is impossible but I wonder if there are any software virtualisation platforms that could be run and that would work. Does anyone know if something like this exists/is possible? Thanks! C

    Read the article

  • AD, Windows-NT Authentication queries

    - by rockbala
    Need Help on the following questions. When a users login (on a computer in the network) is validated against AD what is/are the authentication method used? When a users login is validated in Windows NT environment (not AD) what is/are the authentication method used? If all user's account is on AD, is it possible to change the authentication mechanism only (or protocols) from AD to NT and vice versa (if possible)? If part/whole of question 3 is valid, where should one look to change these authentication methods ? What is the difference between AD and Windows-NT authentication ? Windows server 2008 Domain controller used. Regards, Balaji

    Read the article

  • SQL restore from single file db to filegroup

    - by Mauro
    I have a 180GB MOSS 2007 database whose maintenance (i.e. backups and restores) are becoming a problem. Part of the issue can be resolved by splitting the three content sites down into their own site collections, however this will likely still leave me with a 100gb DB to deal with. Whilst this isnt entirely problematic for SQL it does mean that backups / restores take far too long. my idea is to split each of the databases to 30gb files, then to import the content into them which should distribute the content across the file groups,making it much easier / faster to backup/restore. Is there a way to backup from a single file and restore to a filegroup? If i have the wrong understanding of filegroups then I'm more than happy to find out other methods of managing the size of databases.

    Read the article

< Previous Page | 233 234 235 236 237 238 239 240 241 242 243 244  | Next Page >