Search Results

Search found 24132 results on 966 pages for 'non clustered index'.

Page 405/966 | < Previous Page | 401 402 403 404 405 406 407 408 409 410 411 412  | Next Page >

  • Find rows that match multiple (two) criteria

    - by Scott Sanders
    I am trying to get a list of items that match on two criteria. DATE PRODUCT CHANNEL DEC 13 Cat Chow DRUG NOV 13 Pupperoni DRUG DEC 13 Meow Mix DRUG DEC 13 Jerky Treats DRUG This is the data from A1 to C5. I want all the product sold on DEC 13 and in the DRUG channel. This is the formula that I have now, which is only bringing back the first example. INDEX($B$2:$B$5, MATCH(1, ("DRUG"=$C$2:$C$5)*("DEC 13"=$A$2:$A$5), 0) )

    Read the article

  • Postgres + OpenStreeMap Amazon snapshot

    - by user32425
    Hi, I am trying to start my postgres using /etc/init.d/postgresql-8.3 start but I got this * Starting PostgreSQL 8.3 database server * Error: The server must be started under the locale : which does not exist any more. ...fail! I tried the solution in http://ubuntuforums.org/archive/index.php/t-397005.html but it still does not work. How can I fix this? Thanks

    Read the article

  • What are some useful Windows key shortcut keys for Windows 8 [closed]

    - by mcqwerty
    Possible Duplicate: What are the new shortcuts for Windows 8? There are lots of new UI elements in Windows 8 such as the 'Charms Bar' and the fact that the Desktop is treated as an app. Some elements are easier than others to access on a non-touch screen enabled device such as a standard desktop PC or laptop equipped with a keyboard and mouse. What are some useful shortcut keys that utilise the Windows Key (Win) to make using Windows 8 easier via a keyboard and mouse?

    Read the article

  • Why can't we reach some (but not all) external web service via VPN connection?

    - by Paul Haldane
    At work (UK university) we use a set of Windows servers running WS2008R2 and RRAS which offer VPN service to students in our accommodation. We do this to associate the network connections with individuals. Before they've connected to the VPN all they can talk to is the stuff thats needed to setup the VPN and a local web site with documentation on how to connect. Medium term we'll probably replace this but it's what we're using at the moment. VPN on the 2008 servers allocates client a private (10.x) address. Access to external sites is through NAT on the campus routers (same as any other directly connected client on a private address). Non-VPN connections aren't seeing this problem. Older servers run WS 2003 and ISA2004. That setup works but has become unreliable under load. Big difference there was that we were allocating non-RFC1918 addresses to the clients (so no NAT required). Behaviour we're seeing is that once connected to the VPN, clients can reach local web sites (that is sites on the campus network) but only some external sites. It seems (but this may be chance) that the sites we can reach are Google ones (including YouTube). We certainly have trouble reaching Microsoft's Office 365 service (which is a pain because that's where mail for most of our students is). One odd bit of behaviour is that clients can fetch (using wget on a Windows 7 client) http://www.oracle.com/ (which gets a 301 redirect) but hangs when asked to fetch http://www.oracle.com/index.html (which is what the first URL redirects to). Access works reliably if we configure clients to use our local web proxies (Squid). My gut tells me that this is likely to be something in the chain dropping replies either based on HTTP inspection or the IP address in the reply. However I'm puzzled about why we're seeing this with the VPN clients. Plan for tomorrow (when I'm back in the office) is to setup a web server on external connection so that we can monitor behaviour at both ends of the conversation (hoping that the problem manifests itself with our test server). Any suggestions for things we should be looking at?

    Read the article

  • Reliable file copy (move) process - mostly Unix/Linux

    - by mfinni
    Short story : We have a need for a rock-solid reliable file mover process. We have source directories that are often being written to that we need to move files from. The files come in pairs - a big binary, and a small XML index. We get a CTL file that defines these file bundles. There is a process that operates on the files once they are in the destination directory; that gets rid of them when it's done. Would rsync do the best job, or do we need to get more complex? Long story as follows : We have multiple sources to pull from : one set of directories are on a Windows machine (that does have Cygwin and an SSH daemon), and a whole pile of directories are on a set of SFTP servers (Most of these are also Windows.) Our destinations are a list of directories on AIX servers. We used to use a very reliable Perl script on the Windows/Cygwin machine when it was our only source. However, we're working on getting rid of that machine, and there are other sources now, the SFTP servers, that we cannot presently run our own scripts on. For security reasons, we can't run the copy jobs on our AIX servers - they have no access to the source servers. We currently have a homegrown Java program on a Linux machine that uses SFTP to pull from the various new SFTP source directories, copies to a local tmp directory, verifies that everything is present, then copies that to the AIX machines, and then deletes the files from the source. However, we're finding any number of bugs or poorly-handled error checking. None of us are Java experts, so fixing/improving this may be difficult. Concerns for us are: With a remote source (SFTP), will rsync leave alone any file still being written? Some of these files are large. From reading the docs, it seems like rysnc will be very good about not removing the source until the destination is reliably written. Does anyone have experience confirming or disproving this? Additional info We will be concerned about the ingestion process that operates on the files once they are in the destination directory. We don't want it operating on files while we are in the process of copying them; it waits until the small XML index file is present. Our current copy job are supposed to copy the XML file last. Sometimes the network has problems, sometimes the SFTP source servers crap out on us. Sometimes we typo the config files and a destination directory doesn't exist. We never want to lose a file due to this sort of error. We need good logs If you were presented with this, would you just script up some rsync? Or would you build or buy a tool, and if so, what would it be (or what technologies would it use?) I (and others on my team) are decent with Perl.

    Read the article

  • php mysql cpanel high cpu usage

    - by Megahostzone Santu
    server taking high cpu usage load average: 108.87, 105.92, 85.82 netstat -ntu | awk '{print $5}' | cut -d: -f1 | sort | uniq -c | sort -n Reselt showing too much connect from server IP cpanel Process Manager showing 19.4 | 0.5 | /usr/sbin/mysqld --basedir=/ --datadir=/var/lib/mysql --user=mysql --log-error=/var/lib/mysql/zebra546.serverstall.com.err --pid-file=/var/lib/mysql/zebra546.serverstall.com.pid 3.0 | 0.2 | /usr/bin/php /home/nowwatch/public_html/index.php

    Read the article

  • How to remove Unicode characters and/or convert OpenOffice spreadsheet cells to plaintext?

    - by gonzobrains
    I have an OpenOffice spreadsheet into which I occasionally copy/paste snippets from web pages. However, I need the file, as a whole, to be free of fancy formatting and non-ASCII text. Is tried highlighting cells and selecting "Default Formatting" but this still seems to keep extraneous characters even though it looks like normal text to the human eye. If this is not possible, is there a way to at least reveal the "raw" data within a cell so that I can manually strip it? Thanks, Jeff

    Read the article

  • rsync delete remote duplicates

    - by BlakBat
    I'm trying to delete remote duplicate files without transferring the non-existing files, and without updating the existing files. If I specify both --existing and --ignore-existing (along with "-av --remove-source-files", the operation is a no-op and nothing will be transfered, but nothing will be deleted either. The best I got so far is to make a local copy of destination, use rsync without --ignore-existing, then rsync my local copy on top of the destination

    Read the article

  • What does 'Can't sync file './database_name/…frm' to disk (Errcode: 28)' mean?

    - by cool_cs
    I am getting this error message whenever I try to create a new index or a new table in my mysql server. Does anyone know what the reason is? This is the ouput after I run df -a Filesystem Size Used Avail Use% Mounted on /dev/sda2 13G 7.3G 4.5G 63% / /dev/sda1 251M 27M 212M 12% /boot tmpfs 3.9G 0 3.9G 0% /dev/shm 10.156.248.29:/vol/pharos_pnxd_data_01/env_empty_sbid_27133_qdcprod 30G 30G 32K 100% /app

    Read the article

  • Do I need the meta-key for vim?

    - by Riyaah
    I'm looking in to learning emacs or vim. I started out with emacs but found the need for a meta key to be a hassle, especially since I have a non-english keyboard layout on my macbook. So far I haven't seen any references to meta in vim, so my question is: Can I live without meta in vim? If so that'll settle the vim vs. emacs question for me, otherwise I'll just have to learn to live with some workaround.

    Read the article

  • How to keep multiple servers in sync file wise?

    - by GForceSys
    I'm currently managing a cluster of PHP-FPM servers, all of which tend to get out of sync with each other. The application that I'm using on top of the app servers (Magento) allows for admins to modify various files on the system, but now that the site is in a clustered set up modifying a file only modifies it on a single instance (on one of the app servers) of the various machines in the cluster. Is there an open-source application for Linux that may allow me to keep all of these servers in sync? I have no problem with creating a small VM instance that can listen for changes from machines to sync. In theory, the perfect application would have small clients that run on each machine to be synced, which would talk to the master server which would then decide how/what to sync from each machine. I have already examined the possibilities of running a centralized file server, but unfortunately my app servers are spread out between EC2 and physical machines, which makes this unfeasible. As there are multiple app servers (some of which are dynamically created depending on the load of the site), simply setting up a rsync cron job is not efficient as the cron job would have to be modified on each machine to send files to every other machine in the cluster, and that would just be a whole bunch of unnecessary data transfers/ssh connections.

    Read the article

  • Experiences with eXdupe?

    - by ewwhite
    I noticed that the eXdupe compression/archiving/deduplication utility was recently mentioned in another post here. It boasts some interesting features, and I've been playing with it for the past day. It's basically a cross-platform, highly multithreaded archival tool. http://exdupe.com/index.html I'm curious if anyone here uses it in production or has any tips on how to leverage the tool in their environment. I'm looking for suggestions.

    Read the article

  • Tools for reformatting a spreadsheet into PDF?

    - by Todd
    I'm looking for an easy way to produce nicely formatted PDF documents using a subset of data available in a spreadsheet. Ideally, I would be able to produce one or more PDF documents for each row int e spreadsheet. Each row contains scientific data about a sample. There are numbers, text and images in the Spreadsheet that need to be reformatted in a nice way in a PDF. Ideally, the solution would be a piece of software that a non-technical but highly intelligent scientist could operate.

    Read the article

  • Apache redirects directories

    - by Ziaix
    So, I'm trying to redirect any pages to a file, but avoid redirecting anything thats an existing file or directory. RewriteEngine On RewriteCond %{DOCUMENT_ROOT}%{REQUEST_FILENAME} -d RewriteRule ^(.+)$ $1 [L] RewriteCond %{DOCUMENT_ROOT}%{REQUEST_FILENAME} !-f RewriteRule ^(.+)$ /index.php?page=$1 [QSA] However, any directories still get redirected (existing files are fine and can be located.)

    Read the article

  • Nginx common configuration that I might have missed

    - by ApPeL
    I recently moved from Apache Mod_wsgi to Nginx, and I have seen a major improvement on speed a lowering on memory usage and I am generally very happy with the it. I am not a server expert, so please be gentle. I am wondering if there are any small configuration that I might have missed, that will cause me some issues in the long run... Please see my nginx.conf file user nginx nginx; worker_processes 4; error_log /var/log/nginx/error_log info; events { worker_connections 1024; use epoll; } http { include /etc/nginx/mime.types; default_type application/octet-stream; log_format main '$remote_addr - $remote_user [$time_local] ' '"$request" $status $bytes_sent ' '"$http_referer" "$http_user_agent" ' '"$gzip_ratio"'; client_header_timeout 10m; client_body_timeout 10m; send_timeout 10m; connection_pool_size 256; client_header_buffer_size 1k; large_client_header_buffers 4 2k; request_pool_size 4k; gzip on; gzip_min_length 1100; gzip_buffers 4 8k; gzip_types text/plain; output_buffers 1 32k; postpone_output 1460; sendfile on; tcp_nopush on; tcp_nodelay on; keepalive_timeout 75 20; ignore_invalid_headers on; index index.html; server { listen 80; server_name localhost; location /media/ { root /www/django_test1/myapp; # Notice this is the /media folder that we create above } location /mediaadmin/ { alias /opt/python2.6/lib/python2.6/site-packages/django/contrib/admin/media/; # Notice this is the /media folder that we create above } location / { # host and port to fastcgi server fastcgi_pass 127.0.0.1:8080; fastcgi_param SERVER_ADDR $server_addr; fastcgi_param SERVER_PORT $server_port; fastcgi_param SERVER_NAME $server_name; fastcgi_param SERVER_PROTOCOL $server_protocol; fastcgi_param PATH_INFO $fastcgi_script_name; fastcgi_param REQUEST_METHOD $request_method; fastcgi_param QUERY_STRING $query_string; fastcgi_param CONTENT_TYPE $content_type; fastcgi_param CONTENT_LENGTH $content_length; fastcgi_pass_header Authorization; fastcgi_intercept_errors off; client_max_body_size 100M; } access_log /var/log/nginx/localhost.access_log main; error_log /var/log/nginx/localhost.error_log; } }

    Read the article

  • recommendations for public facing file server

    - by Seaweed
    I've been asked to setup a server for exchanging non-critical company data e.g. marketing PDF's etc. Can anyone recommend a Linux-based solution I can use? We're looking for something with a really simple GUI front end here... I've already setup SFTP but they found the process 'too-techy'. Any ideas would be really welcome... Thanks!

    Read the article

  • How can I instruct nginx to interpret an individual file as php

    - by codecowboy
    A legacy application relies on the following .htaccess rule for an individual file <Files MyFile.html> AddType application/x-httpd-php .html </Files> The new machine has nginx and php-fpm installed with the following config: location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; fastcgi_pass unix:/var/run/php5-fpm.sock; fastcgi_index index.php; include fastcgi_params; }

    Read the article

  • How is network traffic routed (VPN vs. 'open')?

    - by craibuc
    If my workstation has a VPN connection to a given network (a client's, for instance) and an open connection, what determines how a request for a network resource (e.g. a web page) is routed? Moreover, given that a resource (e.g. google.com) could be available via either route (i.e. VPN or non-VPN), how is this route determined? Is there a way to 'force' the routing to use a given route or the route with lower overhead?

    Read the article

  • How to give user read/write access to folders?

    - by Will
    I'm running a certain script that is using a non-root user to do the following... mkdir: cannot create directory `/srv/www/example.com/releases' *** [err :: 12.23.45.789] : Permission denied How would I allow user xyz to have permanent permissions to do so and still keep this web server secure? Also is it possible to make it recursive for all subfolders? I know its probably chmod something but I'm not that linux savy, thanks.

    Read the article

  • Making many network shares appear as one

    - by jimbojw
    Givens: disk is cheap, and there's plenty lying around on various computers around the corporate intranet redundant contiguous large storage volumes are expensive Problem: It would be fantastic to have a single entry point (drive letter, network path) that presents all this space as one contiguous filesystem, effectively abstracting the disk and network architecture from the paths presented to users. Does anyone know how to implement such a solution? I'm open to Windows and non-windows solutions, free and proprietary.

    Read the article

  • Avoiding memory full -> swap full -> crash

    - by Noam
    I'm experiencing an issue when sometimes the memory gets 100% full, and the swap file also, and the server becomes non-responsive and has to be restarted (causing also problems in database). This is what Cacti shows: The server is running a web-app (database + apache) and during that specific moment didn't experience any ir-regular traffic or usage. This scenario happened twice in the last week. What can cause this? How can I resolve the issue?

    Read the article

< Previous Page | 401 402 403 404 405 406 407 408 409 410 411 412  | Next Page >