Search Results

Search found 5521 results on 221 pages for 'deeper understanding'.

Page 52/221 | < Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >

  • books on web server technology

    - by tushar
    i need to understand the web server technologies as to how are the packets recieved how does it respond and understand httpd.conf files and also get to undertand what terms like proxy or reverse proxy actually mean. but i could not find any resources so please help me and suggest some ebook or web site and by server i dont mean a specific one (apache or nginx..) in short a book on understanding the basics about a web server i already asked this on stackoverflow and webmasters in a nutshell was the answer i got and they said its better if i ask it here so please help me out

    Read the article

  • Can I have nginx put stuff on memcached?

    - by adamo
    It is my understanding that nginx can query memcached for data that it serves, but you have to place them there using some other (homegrown) program. So the question is, can I have nginx fetch data from the backend and place them into memcached for future use for certain locations? Sort of not having to place all files from /images in it "by hand" but having them (mem)cached after being fetched for the first time? Any backend other than memcached would suffice also.

    Read the article

  • Why does CPU processing time matter when compared to real wall clock time?

    - by PeanutsMonkey
    I am running the command time 7zr a -mx=9 sample.7z sample.log to gauge how long it takes to compress a file larger than 1GB. The results I get are as follows. real 10m40.156s user 17m38.862s sys 0m5.944s I have a basic understanding of the difference but don't understand how this plays a role in the time in takes to compress the file. For example should I be looking at real or user + sys?

    Read the article

  • Server-Ip address is not getting displayed in snmp-trap messages

    - by Gaurav
    my problem is about Ip Address which I am receiving on the windows system snmp-trap messages,is some thing like that UDP: [192.168.1.150]:1029->[0.0.0.0]:0 while same trap message on linux system has been displayed as UDP: [192.168.1.150]:1030->[192.168.1.23] Now,if you observed carefully these two ,its clearly shown that server ip is not coming in windows traps-Ip([0.0.0.0]:0).what would be the possible reason?please anyone can help me about understanding this & can provide the solution for this problem.

    Read the article

  • Good Booking Engines Suggestions

    - by user28139
    I want to get some suggestion for a Customizable/Opensource Booking Engine for Hotels That you've used or had coded them. The booking engine I'm looking for is the one you can add and customize their fields (address,rates, and other stuffs). Also I can easily integrate it to my existing site. I was looking at CultBooking but I've been having hard time understanding their interface and documentation. Thanks in advance!

    Read the article

  • How do large blobs affect SQL delete performance, and how can I mitigate the impact?

    - by Max Pollack
    I'm currently experiencing a strange issue that my understanding of SQL Server doesn't quite mesh with. We use SQL as our file storage for our internal storage service, and our database has about half a million rows in it. Most of the files (86%) are 1mb or under, but even on fresh copies of our database where we simply populate the table with data for the purposes of a test, it appears that rows with large amounts of data stored in a BLOB frequently cause timeouts when our SQL Server is under load. My understanding of how SQL Server deletes rows is that it's a garbage collection process, i.e. the row is marked as a ghost and the row is later deleted by the ghost cleanup process after the changes are copied to the transaction log. This suggests to me that regardless of the size of the data in the blob, row deletion should be close to instantaneous. However when deleting these rows we are definitely experiencing large numbers of timeouts and astoundingly low performance. In our test data set, its files over 30mb that cause this issue. This is an edge case, we don't frequently encounter these, and even though we're looking into SQL filestream as a solution to some of our problems, we're trying to narrow down where these issues are originating from. We ARE performing our deletes inside of a transaction. We're also performing updates to metadata such as file size stats, but these exist in a separate table away from the file data itself. Hierarchy data is stored in the table that contains the file information. Really, in the end it's not so much what we're doing around the deletes that matters, we just can't find any references to low delete performance on rows that contain a large amount of data in a BLOB. We are trying to determine if this is even an avenue worth exploring, or if it has to be one of our processes around the delete that's causing the issue. Are there any situations in which this could occur? Is it common for a database server to come to the point of complete timeouts when many of these deletes are occurring simultaneously? Is there a way to combat this issue if it exists? (cross-posted from StackOverflow )

    Read the article

  • Will running CSF & Bastille cause any conflicts?

    - by MatW
    I'm taking my first steps into the world of un-managed servers, and have confused myself whilst reading through the 101 tutorials on server hardening that Google spews out! The most recent advice I have been given is to install both CSF and Bastille on my server (used to serve a consumer-facing ecommerce site and act as the business' email server), but my understanding was that both of these tools were an abstraction layer above netfilter / iptables. Will installing both packages cause any conflicts, or do they play well together?

    Read the article

  • How big of a bandwidth hog is Internet radio?

    - by jmgant
    I was thinking about logging into Pandora at work like I do at home, but I'm concerned about sucking up all of the available bandwidth on the network with something that's not strictly work-related. I don't have a thorough technical understanding of how streaming content like Internet radio is delivered, so I don't really know how to measure the impact. Can anyone offer any perspective on how much bandwidth Internet radio consumes relative to normal Internet browsing? Is there any way to measure how much I'm using for a specific site like Pandora?

    Read the article

  • Swapping root directory of a web server

    - by Austin
    One of the clients that I maintain requests a website redesign, we do all the work in a subfolder /newsite/ and they intend to make it live this instant. The /newsite/ is made in wordpress and from my understanding WP keeps all the URL's as they were initially created. Would it be easier to just create the new root dir at /newsite/ or move all contents of the root to /oldsite/ and move /newsite/ to the root? Please advise the best route, i.e. mod_rewrite.

    Read the article

  • Windows 7 memory usage

    - by lydonchandra
    Physical memory(MB) for Windows 7 Total 4021 Cached 1113 Available 768 Free 174 Memory used 3.25GB At this point, windows7 asks me to close some programs because "system memory is low". From my understanding reading articles, I still have 768 MB free memory, why does windows7 complain? Also what does Cached memory refer to? Is this part of memory that Windows7 reserved for itself meaning it's free to use by Windows7 (and means I have about 768 + 1113 MB of free mem?)?

    Read the article

  • How do private BitTorrent trackers monitor how much users upload/download?

    - by Jon-Eric
    From Wikipedia: Most private trackers monitor how much users upload or download, and in most situations, enforce a minimum upload-to-download ratio. How exactly can a tracker figure out how much data was uploaded and downloaded by each user? My understanding is that a BitTorrent tracker is merely a registry of users that are currently downloading/seeding and that peers, once connected, transfer data directly. So I wouldn't think that the tracker would know anything about the amount of data transferred, much less, where it came from.

    Read the article

  • Launch Nginx on startup

    - by Richard Schneeman
    I was able to install and run nginx, and passenger but i'm not able to have nginx come up whenever i reboot my server. To start the nginx process i just type in sudo /opt/nginx/sbin/nginx. From my understanding anything i put in /etc/init.d with ubuntu hardy will be execute. Does anyone know the command that will mimic this command line call? sudo /opt/nginx/sbin/nginx

    Read the article

  • What is the best IIS tracing tool you have used?

    - by Vivek
    I have spent majority of my career using and troubleshooting IIS Web Server. According to me the best thing happened to a Web admin is FRT (Failed Request Tracing) in IIS 7.0.I have used Event Tracing for Windows as well and FRT is as much helpful.Is there any such tracing tool which can give such good in-depth and greater understanding on request flow through the pipeline?

    Read the article

  • Does VMware ESX Fault Tolerance (FT) support depend on the CPU only?

    - by user71784
    I'm trying to find out whether VMware ESX 4.x Fault Tolerance (FT) is supported on a particular server and VMware's HCL is confusing me. It says that some servers with FT-supported processors (specifically the Xeon 3400 Lynnfield) do not support FT and some with almost identical specs (same chipset for instance) do support FT. Could this be a mistake on the HCL itself? To my understanding FT support is based only on the CPU. Thanks. RC

    Read the article

  • Tomcat maximum threads

    - by Clinton Bosch
    I understand that setting the maximum number of connections available in a connection pool should be the same as your maxThreads configured for your Tomcat server (which correlates to the number of requests that can be handled) For tomcat the default is 200, I assume there is a maximum that you can safely configure for your Tomcat server before things start getting out of control, which I assume is also governed by the resources of the machine it is running on. I am trying to get an understanding of the size of maxThreads that people are using with success, is 1000 too big?

    Read the article

  • What causes Mac OS X Permission errors?

    - by Matthew Savage
    This is out of interest rather then looking for a fix to a problem. What actually causes permissions on Mac OS X Systems to become messed up? Its an easily fixed problem (i.e. there's a quick and easy fix via Disk Utility) but its something I'd encountered a few times doing support in a Mac-reseller store without actually understanding the causes. I'd guess that part of it is due to some applications not playing nicely, but what else might be the source of this issue?

    Read the article

  • localhost as hostname confusion [duplicate]

    - by Baboon
    This question already has an answer here: localhost as hostname confusion 1 answer I have a basic understanding about hostname and FQDN. Now I am confused, do I really have to specify a name for my hostname? So for example: Hostname: somename Domain: mydomain.com FQDN: somename.mydomain.com Now, I see something that the hostname is localhost. What is the difference and impact of that? So my FQDN if localhost is my hostname would be localhost.mydomain.com, right?

    Read the article

  • MySQL Backup - incremental

    - by Tiffany Walker
    I know that you can use mysqldump. I am currently dumping the following way: ${MYSQLDUMP} --single-transaction -u ${MUSER} -h ${MHOST} -p${MPASS} $db | ${GZIP} -9 > $FILE From my understanding this locks the database and prevents any type of use of the database and can even lock up websites. Is there a better way to maybe do daily/hourly backups of the MySQL database should the database be in the 100mbs and even 1gbs in size?

    Read the article

  • IIS, SSL with client certs on web farm

    - by Jeremy
    We're building a web service that will be deployed on an IIS 7.5 farm, and secured through SSL, and also requiring client certs that will be mapped to Active Directory accounts. My understanding is that the server cert needs to be generated for a specific server. If that is the case then we will need a server cert for each server in the farm. Because the farm will be load balanced, how do we generate client certs that will work with any of the servers in the farm?

    Read the article

  • Postgresql Internals - Documentation

    - by NogginTheNog
    I'm looking for some up-to-date information about postgresql internals, specifically the query optimizer. I've found this link (referred to in the "Further Reading" section of the 8.4 docs):- http://db.cs.berkeley.edu//papers/UCB-MS-zfong.pdf but it seems quite old. That in itself is not a problem, but I wanted to be sure that I have information that is relevant. Is this the best resource for understanding how postgresql processes queries (using plans, statistics etc.) or are there others?

    Read the article

  • Dns server: how can I configure a dns server to track down a pc name in a dhcp enviroment?

    - by Pitto
    I have a dhcp in my home and I would like to setup a dns server too. I would like to implement a linux solution but I think I can't get hands on without understanding - very superficially - if I can achieve such result. My pc (hostname: test) gets a 192.168.1.7 from dhcp. Its dns server is my router (192.168.1.1). How can the router relate my ip change (as soon as the lease is over) to my hostname?

    Read the article

< Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >