Search Results

Search found 31319 results on 1253 pages for 'source engine'.

Page 630/1253 | < Previous Page | 626 627 628 629 630 631 632 633 634 635 636 637  | Next Page >

  • What else can I do to secure my Linux server?

    - by eric01
    I want to put a web application on my Linux server: I will first explain to you what the web app will do and then I will tell you what I did so far to secure my brand new Linux system. The app will be a classified ads website (like gumtree.co.uk) where users can sell their items, upload images, send to and receive emails from the admin. It will use SSL for some pages. I will need SSH. So far, what I did to secure my stock Ubuntu (latest version) is the following: NOTE: I probably did some things that will prevent the application from doing all its tasks, so please let me know of that. My machine's sole purpose will be hosting the website. (I put numbers as bullet points so you can refer to them more easily) 1) Firewall I installed Uncomplicated Firewall. Deny IN & OUT by default Rules: Allow IN & OUT: HTTP, IMAP, POP3, SMTP, SSH, UDP port 53 (DNS), UDP port 123 (SNTP), SSL, port 443 (the ones I didn't allow were FTP, NFS, Samba, VNC, CUPS) When I install MySQL & Apache, I will open up Port 3306 IN & OUT. 2) Secure the partition in /etc/fstab, I added the following line at the end: tmpfs /dev/shm tmpfs defaults,rw 0 0 Then in console: mount -o remount /dev/shm 3) Secure the kernel In the file /etc/sysctl.conf, there are a few different filters to uncomment. I didn't know which one was relevant to web app hosting. Which one should I activate? They are the following: A) Turn on Source Address Verification in all interfaces to prevent spoofing attacks B) Uncomment the next line to enable packet forwarding for IPv4 C) Uncomment the next line to enable packet forwarding for IPv6 D) Do no accept ICMP redirects (we are not a router) E) Accept ICMP redirects only for gateways listed in our default gateway list F) Do not send ICMP redirects G) Do not accept IP source route packets (we are not a router) H) Log Martian Packets 4) Configure the passwd file Replace "sh" by "false" for all accounts except user account and root. I also did it for the account called sshd. I am not sure whether it will prevent SSH connection (which I want to use) or if it's something else. 5) Configure the shadow file In the console: passwd -l to lock all accounts except user account. 6) Install rkhunter and chkrootkit 7) Install Bum Disabled those services: "High performance mail server", "unreadable (kerneloops)","unreadable (speech-dispatcher)","Restores DNS" (should this one stay on?) 8) Install Apparmor_profiles 9) Install clamav & freshclam (antivirus and update) What did I do wrong and what should I do more to secure this Linux machine? Thanks a lot in advance

    Read the article

  • Cloud storage provider lost my data. How to back up next time?

    - by tomcam
    What do you do when cloud storage fails you? First, some background. A popular cloud storage provider (rhymes with Booger Link) damaged a bunch of my data. Getting it back was an uphill battle with all the usual accusations that it was my fault, etc. Finally I got the data back. Yes, I can back this up with evidence. Idiotically, I stayed with them, so I totally get that the rest of this is on me. The problem had been with a shared folder that works with all 12 computers my business and family use with the service. We'll call that folder the Tragic Briefcase. It is a sort of global folder that's publicly visible to all computers on the service. It's our main repository. Today I decided to deal with some residual effects of the Crash of '11. Part of the damage they did was that in just one of my computers (my primary, of course) all the documents in the Tragic Briefcase were duplicated in the Windows My Documents folder. I finally started deleting them. But guess what. Though they appeared to be duplicated in the file system, removing them from My Documents on the primary PC caused them to disappear from the Tragic Briefcase too. They efficiently disappeared from all the other computers' Tragic Briefcases as well. So now, 21 gigs of files are gone, and of course I don't know which ones. I want to avoid this in the future. Apart from using a different storage provider, the bigger picture is this: how do I back up my cloud data? A complete backup every week or so from web to local storage would cause me to exceed my ISP's bandwidth. Do I need to back up each of my 12 PCs locally? I do use Backupify for my primary Google Docs, but I have been storing taxes, confidential documents, Photoshop source, video source files, and so on using the web service. So it's a lot of data, but I need to keep it safe. Backup locally would also mean 2 backup drives or some kind of RAID per PC, right, because you can't trust a single point of failure? Assuming I move to DropBox or something of its ilk, what is the best way to make sure that if the next cloud storage provider messes up I can restore?

    Read the article

  • Is it possible to copy a set of files, but automatically skip if file already exists?

    - by awe
    I know that the copy command has an option to automatically replace a file if it already exists, but I want to know if it is a way to copy the files only if they not already exist (/Y). I do not know the actual file names in the batch code, as I copy from the source using wildcards in the copy command: copy *.zip c:\destination The reason I want this instead of automatic overwrite is that the files are large, and to skip existing would save a lot of execution time.

    Read the article

  • Need help in using svn on ubuntu 9.10

    - by michael
    Hi, I have install svn on ubuntu 9.10. But when I try to use svn to checkout code for an open source project, i get this error: $ svn co svn://svn.valgrind.org/valgrind/trunk valgrind svn: Berkeley DB error for filesystem '/home/svn/repos/valgrind/db' while opening 'nodes' table: Cannot allocate memory svn: bdb: Lock table is out of available locker entries Can you please tell me how to fix it? Thank you.

    Read the article

  • tmux session switcher now displays windows belonging to sesions

    - by Nick Barnwell
    I recently compiled tmux from source rather than going through homebrew for some reason I can no longer recall. At the same time, switching sessions via ^b sstarted to display all of the windows under each session, and not just the sessions thesmelves: Session Switcher Now Is there any way I can revert to the previous behaviour of only displaying sessions and not the windows belonging to them as well?

    Read the article

  • Is there a way to make IE7 default to ms-image-mode:bicubic?

    - by Mark Ransom
    Internet Explorer 7 uses a rather crude method to resize images by default. There's a CSS tag img { -ms-interpolation-mode: bicubic; } to get better results, but I'm unable to change the source of the page. IE8 uses the better method as the default, but this is a corporate environment that is unable to upgrade at this moment. Is there anything in the options or a registry hack to change the default resizing mode in IE7?

    Read the article

  • Company's logo on Facebook Apps

    - by Iuri Sampaio
    Since a few months ago I've struggling to find a good source explaning how to have my website logo on the facebook apps I attach on my website. In the link bellow (grabbed from facebook website) we can see a fair example of a Share app, with logo, title and description. https://www.facebook.com/sharer/sharer.php?u=https://parse.com I already completed app's profile with all kinds of information the forms have. What do I need to do to show my logo on facebook apps? Best wishes, Iuri

    Read the article

  • Postifix SMTP Load Balance

    - by user103373
    I want to load balance outbound emails between 3 post-fix gateways for sending mails only reason is to use multiple different source IPs to increase throughput & inbox delivery. Each gateway should receive an approximately equal amount of outbound messages. How is it possible please suggest. +---------- smtp A --------- Internet | clients -------- smtp lb ----- smtp B --------- Internet | +---------- smtp C --------- Internet

    Read the article

  • preg_match in .bat file to flusk image url.

    - by marcell22
    welcome, I have problem with .bat script on windows. I use wget to download html stats page, now i have to find (in html source) url like this http://www.example.com/stats/367895.jpeg The 367895 is a random generated number. and download chart jpeg. I think i can't do this in .bat, Do You know any external command line application what i could trigger from .bat and show in output finded url ? Regards

    Read the article

  • SQL Server 2008 hosting (for development)

    - by hazimdikenli
    Hello, We are doing distributed development, working at home, office and sometimes at customers. We are using assembla for source-repository and we need a centralized-remote SQL Server 2008 database hosting for (similar to svn on assembla) our SQL development server. Can you name / recommend any service providers?

    Read the article

  • PHP product recommendations

    - by David
    I'm building up an online store written in PHP. I was wondering if anyone has any suggestions for open source (or otherwise) product recommendation plugins? Ideally I'd like something that stores all its information locally on my server, and can provide good recommendations of products based on previous purchases and clicks. Any hints appreciated!

    Read the article

  • A free WebDAV client for Windows?

    - by Brian L.
    I'd like to copy files from a network drive to a SharePoint site (perhaps as a mapped drive). What's a good client to do so? Obviously Windows (XP) Explorer is bad, I"m trying RichCopy at the moment. Any opinions on CoreFTP? Are there any recommended open-source WebDAV clients?

    Read the article

  • Is there a way to replicate a very large file shares in real-time?

    - by fsckin
    I have an hourly cron job that copies about 40GB of data from a source folder into a new folder with the hour appended on the end. When it's done, the job prunes anything older than 24 hours. This data changes very often during work hours and is on a samba file share. Here's how the folder structure looks: \server\Version.1 \server\Version.2 \server\Version.3 ... \server\Version.24 The contents of each new folder compared to the last one usually doesn't change very much, since this is a hourly job. Now you might be thinking that I'm an idiot for setting dreaming this up. Truth is, I just found out. It's actually been used for years and is so incredibly simple, anyone could delete the ENTIRE 40GB share (imagine that dialog spooling up... deleting thousands and thousands of files) and it would actually be faster to restore by moving the latest copy back to the source than it took to delete. Brilliant! Now to top this off, I need to efficiently replicate this 960GB of "mostly similar" data to a remote server over WAN link, with the replication happening as close to real-time as possible -- think hot spare, disaster recovery, etc. My first thought was rsync. Total failure. Rsync sees it sees a deletion of the folder that is 24 hours old and the addition of a new folder with 30GB of data to sync! I also looked at rdiff-backup and unison, they both appear to use similar algorithms and do not keep enough meta-data to do this intelligently. Best thing that I can find "out of the box" to do this is Windows Server "Distributed Filesystem Replication" which uses "Remote Differential Compression" -- After reading the background information on how this works, it actually looks like exactly what I need. Problem: Both servers are running Linux. D'oh! One approach to this I'm looking at is this, say it's 5AM and the cron job finishes: New Version.5 folder arrives at on local server SSH to remote server and copy Version.4 to Version.5 Run rsync on the local server pushing changes to the remote server. Rsync finally knows to do a differential copy between Version.4 and Version.5 Is there a smarter way to replicate Samba shares as close to real-time as possible? Anything out there that does "Remote Differential Compression" on Linux?

    Read the article

  • Can I stop SharePoint from prompting during file edit?

    - by uSlackr
    We use a SharePoint 2007 site internally with Office 2010. Whenever I open a Word document to edit it, I get a prompt saying: Some files can harm your computer. If the file information below looks suspicious, or you do not fully trust the source, do not open the file. I've been unable to find a reliable answer around the web. Some suggested using the Windows File Types dialog to remove the prompt on download option, but this dialog is not available in Windows 7.

    Read the article

  • Untar multiple files in different directories

    - by wildeep
    Hi, am trying to perform an update on multiple site that use an open source CMS but untaring a patch file in each sites httpdocs directory. My plan was to perform a find for the patch file then untar using the following command: find . -name "patchfile.tar.gz" -exec tar -xzvf {} \; -print but it doesnt seem to work successfully anyone have any ideas as to why not? Many thanks.

    Read the article

  • rsync delete remote duplicates

    - by BlakBat
    I'm trying to delete remote duplicate files without transferring the non-existing files, and without updating the existing files. If I specify both --existing and --ignore-existing (along with "-av --remove-source-files", the operation is a no-op and nothing will be transfered, but nothing will be deleted either. The best I got so far is to make a local copy of destination, use rsync without --ignore-existing, then rsync my local copy on top of the destination

    Read the article

  • Bridging and iptables SNAT conflict

    - by sad_admin
    Hello I am working on a setup here and have it working with one minor exception. Devices on one side of my bridge aren't getting SNAT'd to the Internet. The Diagram / Overview: Primary_Network (Site_A) | | Internet ------- Linux_Bridge_GW (GW) | | Secondary/CoLo Site (Site_B) Here is the setup: 1.) Site_A has all the production servers and workstations. 2.) Site_B has a set of servers that we would like to fail-over to and also serve our internet facing services from. 3.) GW has two interfaces that are trunked and carrying the appropriate VLAN traffic (allow layer-2 propagation of traffic between sites) //this all works perfectly fine. 4.) The problem that is being encountered is, hosts from Site_B have their default GW at Site_A (same subnet) GW does not have IPs on the VLANs that are being passed. 5.) All hosts at Site_A can reach the Internet without problem. 6.) GW has an addresses on a subnet that is ONLY for Internet destined traffic. (This was done so that Websense would not have to parse unnecessary traffic. We use this VLAN as the monitor port's source on the switch where Websense is sitting). What I think is happening: 1.) Packet/Frame comes in on physdev at Site_B destined for Internet. 2.) Kernel sees packet, and forwards it out the other side of the bridge to that host's default GW. 3.) Site_A (containing core-network's Default-GW) sees that packet is destined for a host it doesn't know about, so it sends it to it's default GW (the linux bridge, since it's Internet bound). 4.) The kernel says "Hey, I've seen you before" and therefore doesn't do SNAT'ing on the packet and sends it out to the Internet where it's black-holed. Why I think it's happening: 1.) A tcpdump on the internet facing NIC shows the packet leaving the interface with the private address as it's source. What I would like: 1.) Have the packet SNAT'd. 2.) Something like the below would be awesome a.) packet comes in from Site_B b.) kernel sees that the packet is NOT destined for itself or any private address c.) kernel says "OK, well since you're destined for the Internet I'm going to send you out this interface rather than forward you to your normal default GW that's WAAAY over there." d.) packet comes in from internet and is sent out the appropriate bridge physdev depending on which site the host it's destined for is at. Thanks for any assistance or guidance that you are willing to offer. Best Regards, Sad Admin

    Read the article

  • define variable in linux that can be access in php

    - by sweb
    I add a variable in whole linux varibale in /etc/profile export MYNAME="My Value" how can i access this value in php source code during run via apache web server? in $_SERVER this value doesn't exist. just this keys appear on $_ENV: _ENV["APACHE_RUN_DIR"] /var/run/apache2 _ENV["APACHE_PID_FILE"] /var/run/apache2.pid _ENV["PATH"] /usr/local/bin:/usr/bin:/bin _ENV["APACHE_LOCK_DIR"] /var/lock/apache2 _ENV["LANG"] C _ENV["APACHE_RUN_USER"] www-data _ENV["APACHE_RUN_GROUP"] www-data _ENV["APACHE_LOG_DIR"] /var/log/apache2 _ENV["PWD"] /

    Read the article

  • Alternative to Dropbox (on my server)?

    - by jweede
    I love using Dropbox to sync files between all my machines, and I've heard it uses rsync internally to keep files synced. Sometimes I need to sync very large things, and I don't necessarily want to pay for storage space on someone else's server when I have my own. So does anyone know of any nice cross-platform (pref. open source) automatic file-sync applications out there for this? sidenote: Here is a Dropbox referral link, if you're feeling generous.

    Read the article

  • I added __git_ps1 to my PS1 in .bash_profile, now I'm getting (master) for all folders that aren't git repos.

    - by Matthew
    I'm on a Mac (10.6.5). Here's an example of what's going wrong: [m@m ~ (master)]$ cd ~/Documents [m@m ~/Documents (master)]$ cd ~/Applications [m@m ~/Applications (master)]$ cd ~/Library [m@m ~/Library (master)]$ cd ~/Sites/somesite [m@m ~/Sites/somerepo (FEATURE_SOMEFEATURE)]$ Here's the relevant contents of my .bash_profile: source ~/.git-completion.bash PS1='[\u@\h \w$(__git_ps1 " (%s)")]\$ ' I'm using the standard git-completion script - I just copied it to my home directory.

    Read the article

  • Where is the best place to label a server?

    - by Celeritas
    There isn't much room on server chassis and I'm wondering where a label with the servers name should go? Is there any other information in addition to the name that should go on the label? Does it make sense to label each hard drive in a server or is that not necessary? There certainly is overkill, when I worked at Big Blue labeling was a huge source of bureaucracy, even a projector needed to be labelled and have it's whereabouts routinely reported.

    Read the article

  • linux cron job error

    - by bell
    I have setup a cron job to run a php file every 30 minutes, lynx -source public_html/scripts/file.php the result comes through to an email but seems to get this error Can't Access `file://localhost/home/username/public_html/scripts/file.php' Alert!: Unable to access document. lynx: Can't access startfile any advice would be much appreciated

    Read the article

< Previous Page | 626 627 628 629 630 631 632 633 634 635 636 637  | Next Page >