Search Results

Search found 21061 results on 843 pages for 'bulid process'.

Page 363/843 | < Previous Page | 359 360 361 362 363 364 365 366 367 368 369 370  | Next Page >

  • In-memory DB to perform intersects on set slices

    - by IanC
    I have a specific programming need where I need to efficiently store large sorted sets in memory, query them for ranges, and intersect them against other sets that are also queried for ranged. I am looking at Redis, but I can't see a range slice command. MongoDB can only use 1 index, so it has to perform row-level scans, whereas I wish to process using columns that are intersected. I'm also looking at Counchbase, but can't easily determine from the documentation if it is suited to this. I know it uses Memcached, which is AFAIK not suited to this usage. Could anyone share potential solutions for this specific problem? EDIT For example, I need to perform the following: Get the IDs of all cars where the price is between 2000 and 3000, and intersect that will all cars where the engine capacity is between 3000 and 4000.

    Read the article

  • Implementing cache system in Java Web Application

    - by TGM
    I worked with JPA (Eclipselink implementation) and Hibernate. As I understand these two have great caching systems. I am interested in caching in a Web application and in order to better understand the process I'm trying to implement something on my own. Sadly, I cannot find any in depth documentation about this subject. I'm interested in things like high scalability, sharing memory on different machines and other important theoretical matters. Is there any tutorial or open project I could check out? Thank you! *LE: * I want to cache DB information in POJOs just like JPA or Eclipselink

    Read the article

  • Cannot upgrade system because of this error

    - by user292375
    Setting up mongodb-org-server (2.6.1) ... * Starting database mongod [fail] invoke-rc.d: initscript mongod, action "start" failed. dpkg: error processing mongodb-org-server (--configure): subprocess installed post-installation script returned error exit status 1 No apport report written because MaxReports is reached already dpkg: dependency problems prevent configuration of mongodb-org: mongodb-org depends on mongodb-org-server; however: Package mongodb-org-server is not configured yet. dpkg: error processing mongodb-org (--configure): dependency problems - leaving unconfigured No apport report written because MaxReports is reached already Errors were encountered while processing: mongodb-org-server mongodb-org E: Sub-process /usr/bin/dpkg returned an error code (1)

    Read the article

  • Monitor screen size and programming ease

    - by rrazd
    I recently heard that a big part in successful/quick debugging and easing the process of programming is to use a big screen. I may be purchasing a new computer in the future and this has me wondering: 1)Is the aforementioned statement actually true or is it a bit of a stretch? 2)Have you noticed that this plays a significant enough part to buy a bigger screen if the bigger screen is significantly more expensive? 3)Is it common for developers to work on 13'' laptop screen as their main (and only) workstation (this is what I currently develop on) or is this actually disadvantageous? This may be subjective but any professional opinions/experiences would be greatly appreciated!

    Read the article

  • Plagued by multithreaded bugs

    - by koncurrency
    On my new team that I manage, the majority of our code is platform, TCP socket, and http networking code. All C++. Most of it originated from other developers that have left the team. The current developers on the team are very smart, but mostly junior in terms of experience. Our biggest problem: multi-threaded concurrency bugs. Most of our class libraries are written to be asynchronous by use of some thread pool classes. Methods on the class libraries often enqueue long running taks onto the thread pool from one thread and then the callback methods of that class get invoked on a different thread. As a result, we have a lot of edge case bugs involving incorrect threading assumptions. This results in subtle bugs that go beyond just having critical sections and locks to guard against concurrency issues. What makes these problems even harder is that the attempts to fix are often incorrect. Some mistakes I've observed the team attempting (or within the legacy code itself) includes something like the following: Common mistake #1 - Fixing concurrency issue by just put a lock around the shared data, but forgetting about what happens when methods don't get called in an expected order. Here's a very simple example: void Foo::OnHttpRequestComplete(statuscode status) { m_pBar->DoSomethingImportant(status); } void Foo::Shutdown() { m_pBar->Cleanup(); delete m_pBar; m_pBar=nullptr; } So now we have a bug in which Shutdown could get called while OnHttpNetworkRequestComplete is occuring on. A tester finds the bug, captures the crash dump, and assigns the bug to a developer. He in turn fixes the bug like this. void Foo::OnHttpRequestComplete(statuscode status) { AutoLock lock(m_cs); m_pBar->DoSomethingImportant(status); } void Foo::Shutdown() { AutoLock lock(m_cs); m_pBar->Cleanup(); delete m_pBar; m_pBar=nullptr; } The above fix looks good until you realize there's an even more subtle edge case. What happens if Shutdown gets called before OnHttpRequestComplete gets called back? The real world examples my team has are even more complex, and the edge cases are even harder to spot during the code review process. Common Mistake #2 - fixing deadlock issues by blindly exiting the lock, wait for the other thread to finish, then re-enter the lock - but without handling the case that the object just got updated by the other thread! Common Mistake #3 - Even though the objects are reference counted, the shutdown sequence "releases" it's pointer. But forgets to wait for the thread that is still running to release it's instance. As such, components are shutdown cleanly, then spurious or late callbacks are invoked on an object in an state not expecting any more calls. There are other edge cases, but the bottom line is this: Multithreaded programming is just plain hard, even for smart people. As I catch these mistakes, I spend time discussing the errors with each developer on developing a more appropriate fix. But I suspect they are often confused on how to solve each issue because of the enormous amount of legacy code that the "right" fix will involve touching. We're going to be shipping soon, and I'm sure the patches we're applying will hold for the upcoming release. Afterwards, we're going to have some time to improve the code base and refactor where needed. We won't have time to just re-write everything. And the majority of the code isn't all that bad. But I'm looking to refactor code such that threading issues can be avoided altogether. One approach I am considering is this. For each significant platform feature, have a dedicated single thread where all events and network callbacks get marshalled onto. Similar to COM apartment threading in Windows with use of a message loop. Long blocking operations could still get dispatched to a work pool thread, but the completion callback is invoked on on the component's thread. Components could possibly even share the same thread. Then all the class libraries running inside the thread can be written under the assumption of a single threaded world. Before I go down that path, I am also very interested if there are other standard techniques or design patterns for dealing with multithreaded issues. And I have to emphasize - something beyond a book that describes the basics of mutexes and semaphores. What do you think? I am also interested in any other approaches to take towards a refactoring process. Including any of the following: Literature or papers on design patterns around threads. Something beyond an introduction to mutexes and semaphores. We don't need massive parallelism either, just ways to design an object model so as to handle asynchronous events from other threads correctly. Ways to diagram the threading of various components, so that it will be easy to study and evolve solutions for. (That is, a UML equivalent for discussing threads across objects and classes) Educating your development team on the issues with multithreaded code. What would you do?

    Read the article

  • Which MIME type to compress? and what If I omit the `type` attribute from the HTML?

    - by rockyraw
    Per my request, my webhost had turned mod_deflate ON. In my Cpanel I now have an "Optimize Website" button. Inside that menu I could either choose: "Compress all content" or "Compress the specified MIME types" with the following default MIME types: "text/html text/plain text/xml" Which option should I choose and why? If I choose option 2, which types should I add (is there a recommended list with the exact way they should be written)? According to Google recommendations, I have omitted the type="text/css" attributes from all CSS references, as well as the type="text/javascript" attributes from all script references. Would this hinder the "gzipping" process?

    Read the article

  • VBA functionality in Word 2007 [closed]

    - by Mac
    I have a VB system that utilises VBA and MSWord 20078. I have a few problems that I am hoping you solve for me. I need to search for a string in a Word 2007 document and then pickup all characters into a VB Variable from that string to the next carriage return inclusive. I will then manipulate the contents of the VB variable. When I find the search string I need to know what section of the document I am in. Once I have these two pieces of information I then need to continue to the next occurrence of the search string and repeat my functionality. Once I have dealt with the last occurrence of the search string I need to know that so that I can exit the process. Any assistance would be greatly appreciated

    Read the article

  • design for interruptable operations

    - by tpaksu
    I couldn't find a better topic but here it is; 1) When user clicks a button, code starts t work, 2) When another button is clicked, it would stop doing whatever it does and start to run the second button's code, 3) Or with not user interaction, an electrical power down detected from a connected device, so our software would cancel the current event and start doing the power down procedure. How is this design mostly applied to code? I mean "stop what you are doing" part? If you would say events, event handlers etc. how do you bind a condition to the event? and how do you tell the program without using laddered if's to end it's process? method1(); if (powerdown) return; method2(); if (powerdown) return; etc.

    Read the article

  • Is asking for control totals on a file an outdated means of verifying a file?

    - by CTKeane
    I'm in a new position where I need to process a flat files on a regular basis. The last time I did this was 5 or 6 years ago but as part of the file layout I received control totals. It gave me simplistic information on the file like the total number of records as well as sums of the important fields. This helped me during testing then also during production to verify the file arrived and has correct information. I have asked for similar data for this new project and have hit a wall of no. Is this no longer a standard practice? Is there a better way?

    Read the article

  • Generating CMakeLists.txt

    - by vanna
    I got a bunch of C++ sources files and headers. They may use external libraries such as Boost e.g. I am interested in the process of building binaries for Windows and *nix. Makefiles (*nix) and .vcproj (Windows) call compilers with some specifications such as the order of compilation, compilation options and stuff. CMakeLists.txt can be used by CMake to build either makefiles or .vcproj and use very helpful commands such as recursive search of files, automatic linkage with known libraries, installers, variables that can be used in source files... Is there any existing tool that would generate a CMakeLists.txt from specified options ? Options could be like : scan this folder and make a library out of it, then scan this other folder and make an executable and automatically link both with Boost as well along with a user friendly installer with generated INSTALL.txt and README.txt. Something very powerful like that.

    Read the article

  • gnome-panel has multiple rows -- how to undo that?

    - by towi
    On Ubuntu 12.04 I use openbox with a single "gnome-panel" started. But after trying to configure my two screen in Xinerama mode I got additional rows of panels each time X restarted. And despite I completely reverted to my one-screen-X-config the gnome-panel displayed has 4 rows (4 at top sreen, 4 at bottom) only the outermost row has content, the other 3 are empty there is only one gnome-panel process running some widgets appear 4 times, too, although they run only once when I switch away from openbox using unity I am back to 1 row temporarily What went wrong here? How do I undo that? Things that did not help: rm -Rf .gnome2 .gconf/ .compiz-1/ pkill panel, and restarting it

    Read the article

  • Java : fusion approuvée des éditions SE/EE/ME, la gouvernance du langage remaniée, vers plus d'équité ?

    Java : fusion approuvée des comités exécutifs des éditions SE/EE/ME La gouvernance du langage remaniée, vers plus d'équité ? [IMG]http://adiguba.developpez.com/blog/javacup.png[/IMG] Avec quatorze votes favorables et une abstention de Google du comité exécutif (EC) de Java SE/EE, seize Oui et deux abstentions de RIM et AT&T du comité de Java ME, la fusion des comités exécutifs de Java SE (Standard Edition), Java EE (Entreprise Edition) et Java ME (Micro Edition) vient d'être approuvé. Le JCP (Java Community Process) Document passe ainsi à la version 2.9. Le comité exécutif est un groupe de membres qui guide l'évolution de la technologie Java au sein du JCP. Il...

    Read the article

  • Is there any way to keep launchers synchronized among machines?

    - by MMA
    Most of us use more than one machine for our purpose. The scenario for me is to use one at home, another at office. When comes to the launcher at left, the applications get added in chronological order. You can click and drag an application in launcher, but this is a tedious process. If your application placements are different among the machines, you loose the intuitive mouse click. You need to spend a few seconds to find out where actually the launcher is before you click on it. Is there any way I can keep the launcher applications (at least the locked ones) synchronized among machines? Thanks.

    Read the article

  • How to get my windows docs ; pics, mp3, and docs off an external usb hard drive (sata 3.5 enclosure)

    - by Alan
    Back when I had windows 7, I cloned my internal hard drive to an enclosed external usb hard drive. I then formatted my internal hard drive and installed ubuntu 11.10 as my only OS. How do I migrate my files, pics, mp3's, and etc off the external drive and back to my internal drive which now only has ubuntu on it. (and why is this process not made easier?) I have tried logging in as the root user but I cannot find the external device. I have downloaded several different apps to manage files, usb sticks and etc. PLEASE HELP!!!

    Read the article

  • Can't Reboot into Windows 8.1 from Ubuntu

    - by Extended Range
    I am newb in Ubuntu. Currently I am using an Acer V11 laptop with a pre-intalled Windows 8.1. I follow some guideline (http://www.everydaylinuxuser.com/2014/05/install-ubuntu-1404-alongside-windows.html) which successfully install Ubuntu 14.04 LTS alongside Windows 8.1 with UEFI enabled all the way throughout the installation and my current usage and a GRUB. But I got into some problem. The dual boot mostly works fine as I am able to successfully boot into either one of the Windows 8.1 or Ubuntu when stating my laptop. However it ran into issue after I start the device and use Ubuntu for a while (2 hour for instance) and then restart it: If I press the restart from Ubuntu, the grub still shows and I am still able to choose the Windows Boot Manager. But the Windows loading process was stuck at the Acer Loading Screen and I am not able to see the standard circular Win8 loading progress bar. Does someone know why would this happens?

    Read the article

  • How come many of the AppShowdown apps aren't available in Ubuntu anymore?

    - by kermit666
    After upgrading to 12.10 I've noticed that I can't install some of the nice apps created for the AppShowdown, such as: Cuttlefish Blubphone (Lightread also came out quite late in 12.10) It seems such a waste having these great new apps added to the repository, only to exclude them in the first next version. I'm wondering why aren't these apps automatically available in a newer version of Ubuntu. Is it simply that the API is so different that it requires major rewrites and programmer activity or is it some bureaucratic reason? Are there any plans to improve this process?

    Read the article

  • Decrementing/Incrementing loop variable inside for loop. Is this code smell?

    - by FairDune
    I have to read lines from a text file in sequential order. The file is a custom text format that contains sections. If some sections are out of order, I would like to look for the starting of the next valid section and continue processing. Currently, I have some code that looks like this: for (int currentLineIndex=0; currentLineIndex < lines.Count; currentLineIndex++ ) { //Process section here if( out_of_order_condition ) { currentLineIndex--;//Stay on the same line in the next iteration because this line may be the start of a valid section. continue; } } Is this code smell?

    Read the article

  • Incorrect Google Profile Results

    - by Mikhel
    Where do Google gets the info they display along with your profile in their "Profile Results" that is shown once you hit search for a particular person? When Google shows incorrect info, there's a feedback option at the bottom and it gives you an option to report a problem on which info is wrong. However, this is a tedious task since I've already done this before and they have corrected it but then when I search again after a couple of months, wrong info is shown again. So maybe I need to correct or do something on the process of putting those info into the profile... Not sure where Google gets all these info.. Help?

    Read the article

  • Designing a hierarchical structure with lots of reads and writes?

    - by JD01
    I am in the process of working on a video on demand system part of it involves the management of a hierarchical tree structure (think windows explorer) which allows users to upload videos, move folders around, create new folders etc. User action will be allowed depending on which permissions they have. In addition to managing the structure (creating folders and uploading videos), subscribers will be viewing content (read access). The number of reads will be significantly more than the writes. My question (and it is big one) is should I store the data in a database (for the writes) and also have some sort of cache which will be used for the reads? or do I use two databases? or is there a better solution? Also I will have to resolve concurrency issues which I think optimistic locking on the database will resolve. I have a fair bit about CQRS over the last few months but not sure if this is the way to go. Any ideas?

    Read the article

  • How do I get sleep and hibernate working on an Asus K53z?

    - by RivMac
    Help! New ASUS K53z laptop that has a fresh Ubuntu 12.04 install... However, sleep doesn't work... Tried to follow the manual process in this thread... http://ubuntuforums.org/showthread.php?t=1460790 The dropbox package was a 404, and my results are working... closing the lid of the computer shuts the wireless off, and the screen remains lit up. I'm a relative novice, but have been trying to work in terminal and follow instructions, but with this one I'm having little success.

    Read the article

  • Domain name similar to an other existing one, bad for SEO?

    - by qqfr2507
    I am in the process of choosing a domain name for a personal project. I have found a very good one (let's say it is "myproject.com") but it is very close to another existing domain name ("smyproject.com"). Only the first letter is different. This website has a very different activity from mine. My question is: is it bad for SEO? When someone will type "myproject" in a search engine, is there a risk that the first result will be "smyproject.com" if this website has better SEO than mine? Thanks for your help!

    Read the article

  • Html Element search in Firebug 13 !

    - by Anirudha
    Originally posted on: http://geekswithblogs.net/anirugu/archive/2013/10/26/html-element-search-in-firebug-13.aspxFirebug 13 are currently in development process. In firebug 13 you can search the element the way we do in CSS. This is much better and save time. Now just put your selector in search bar and now you are moving to your code.   In this picture I demo how I search a element of stackoverflow.com this is example of how this feature will work. If you want to use Firebug 13 then follow this blog post to get it https://blog.getfirebug.com/2013/10/25/firebug-1-13-alpha-4/

    Read the article

  • How to get started on working with WebTrends locally using a DEV account access? [on hold]

    - by Knowledge Craving
    I'm a newbie developer on WebTrends Analytics, although I have worked extensively on Adobe SiteCatalyst. Currently I want to integrate WebTrends with the web pages of a local project on my local JBoss EAP application server, so that I can run it locally and see how WebTrends work, along with how the Analytics reports show up. Later on, I may customize the reports data as per the project requirements. However, I can't find anywhere how I can start working on a local development process using a DEV account access. I'm mainly trying to get data on number of user visits per day, as of now, which will cater to multichannel experience. And later on, I will start pulling more data based on targeting and segmentation. Can anyone of you please help and guide me in the right direction? Any help is highly appreciated. Thank you!

    Read the article

  • ?Oracle Database 12c????TTnn TMON??redo??????

    - by Liu Maclean(???)
    ?Oracle 11g? Data Guard?redo?????????3???????: ARCi (FAL – archived redo shipping, ping, local only archivals) NSAi (async) 12.1 name: TTnn , NSSi (sync) –– live redo shipping ????12c?? ??TTnn  ??TT00?????async ???redo??? ???????TMON????Redo transport monitor? SQL> select banner from v$version where rownum=1; BANNER -------------------------------------------------------------------------------- Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production SQL> select program,pid from v$process where program like '%TMON%' or Program like '%TT%'; PROGRAM PID ------------------------------ ---------- ORACLE.EXE (TMON) 7 ORACLE.EXE (TT00) 24 ??????? ?11g ???NSAi async redo ship?????????LGWR?????????,?????redo ????; ???12c?TTnn???redo???????LGWR? ???  ??????async redo ship ??redo??! 11g?: 12c?

    Read the article

  • Async ignored on AJAX requests on Nginx server

    - by eComEvo
    Despite sending an async request to the server over AJAX, the server will not respond until the previous unrelated request has finished. The following code is only broken in this way on Nginx, but runs perfectly on Apache. This call will start a background process and it waits for it to complete so it can display the final result. $.ajax({ type: 'GET', async: true, url: $(this).data('route'), data: $('input[name=data]').val(), dataType: 'json', success: function (data) { /* do stuff */} error: function (data) { /* handle errors */} }); The below is called after the above, which on Apache requires 100ms to execute and repeats itself, showing progress for data being written in the background: checkStatusInterval = setInterval(function () { $.ajax({ type: 'GET', async: false, cache: false, url: '/process-status?process=' + currentElement.attr('id'), dataType: 'json', success: function (data) { /* update progress bar and status message */ } }); }, 1000); Unfortunately, when this script is run from nginx, the above progress request never even finishes a single request until the first AJAX request that sent the data is done. If I change the async to TRUE in the above, it executes one every interval, but none of them complete until that very first AJAX request finishes. Here is the main nginx conf file: #user nobody; worker_processes 1; #error_log logs/error.log; #error_log logs/error.log notice; #error_log logs/error.log info; #pid logs/nginx.pid; events { worker_connections 1024; } http { include mime.types; default_type application/octet-stream; server_names_hash_bucket_size 64; # configure temporary paths # nginx is started with param -p, setting nginx path to serverpack installdir fastcgi_temp_path temp/fastcgi; uwsgi_temp_path temp/uwsgi; scgi_temp_path temp/scgi; client_body_temp_path temp/client-body 1 2; proxy_temp_path temp/proxy; log_format main '$remote_addr - $remote_user [$time_local] "$request" ' '$status $body_bytes_sent "$http_referer" ' '"$http_user_agent" "$http_x_forwarded_for"'; #access_log logs/access.log main; # Sendfile copies data between one FD and other from within the kernel. # More efficient than read() + write(), since the requires transferring data to and from the user space. sendfile on; # Tcp_nopush causes nginx to attempt to send its HTTP response head in one packet, # instead of using partial frames. This is useful for prepending headers before calling sendfile, # or for throughput optimization. tcp_nopush on; # don't buffer data-sends (disable Nagle algorithm). Good for sending frequent small bursts of data in real time. tcp_nodelay on; types_hash_max_size 2048; # Timeout for keep-alive connections. Server will close connections after this time. keepalive_timeout 90; # Number of requests a client can make over the keep-alive connection. This is set high for testing. keepalive_requests 100000; # allow the server to close the connection after a client stops responding. Frees up socket-associated memory. reset_timedout_connection on; # send the client a "request timed out" if the body is not loaded by this time. Default 60. client_header_timeout 20; client_body_timeout 60; # If the client stops reading data, free up the stale client connection after this much time. Default 60. send_timeout 60; # Size Limits client_body_buffer_size 64k; client_header_buffer_size 4k; client_max_body_size 8M; # FastCGI fastcgi_connect_timeout 60; fastcgi_send_timeout 120; fastcgi_read_timeout 300; # default: 60 secs; when step debugging with XDEBUG, you need to increase this value fastcgi_buffer_size 64k; fastcgi_buffers 4 64k; fastcgi_busy_buffers_size 128k; fastcgi_temp_file_write_size 128k; # Caches information about open FDs, freqently accessed files. open_file_cache max=200000 inactive=20s; open_file_cache_valid 30s; open_file_cache_min_uses 2; open_file_cache_errors on; # Turn on gzip output compression to save bandwidth. # http://wiki.nginx.org/HttpGzipModule gzip on; gzip_disable "MSIE [1-6]\.(?!.*SV1)"; gzip_http_version 1.1; gzip_vary on; gzip_proxied any; #gzip_proxied expired no-cache no-store private auth; gzip_comp_level 6; gzip_buffers 16 8k; gzip_types text/plain text/css application/json application/x-javascript text/xml application/xml application/xml+rss text/javascript application/javascript; # show all files and folders autoindex on; server { # access from localhost only listen 127.0.0.1:80; server_name localhost; root www; # the following default "catch-all" configuration, allows access to the server from outside. # please ensure your firewall allows access to tcp/port 80. check your "skype" config. # listen 80; # server_name _; log_not_found off; charset utf-8; access_log logs/access.log main; # handle files in the root path /www location / { index index.php index.html index.htm; } #error_page 404 /404.html; # redirect server error pages to the static page /50x.html # error_page 500 502 503 504 /50x.html; location = /50x.html { root www; } # pass the PHP scripts to FastCGI server listening on 127.0.0.1:9100 # location ~ \.php$ { try_files $uri =404; fastcgi_pass 127.0.0.1:9100; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include fastcgi_params; } # add expire headers location ~* ^.+.(gif|ico|jpg|jpeg|png|flv|swf|pdf|mp3|mp4|xml|txt|js|css)$ { expires 30d; } # deny access to .htaccess files (if Apache's document root concurs with nginx's one) # deny access to git & svn repositories location ~ /(\.ht|\.git|\.svn) { deny all; } } # include config files of "enabled" domains include domains-enabled/*.conf; } Here is the enabled domain conf file: access_log off; access_log C:/server/www/test.dev/logs/access.log; error_log C:/server/www/test.dev/logs/error.log; # HTTP Server server { listen 127.0.0.1:80; server_name test.dev; root C:/server/www/test.dev/public; index index.php; rewrite_log on; default_type application/octet-stream; #include /etc/nginx/mime.types; # Include common configurations. include domains-common/location.conf; } # HTTPS server server { listen 443 ssl; server_name test.dev; root C:/server/www/test.dev/public; index index.php; rewrite_log on; default_type application/octet-stream; #include /etc/nginx/mime.types; # Include common configurations. include domains-common/location.conf; include domains-common/ssl.conf; } Contents of ssl.conf: # OpenSSL for HTTPS connections. ssl on; ssl_certificate C:/server/bin/openssl/certs/cert.pem; ssl_certificate_key C:/server/bin/openssl/certs/cert.key; ssl_session_timeout 5m; ssl_protocols SSLv3 TLSv1 TLSv1.1 TLSv1.2; ssl_ciphers HIGH:!aNULL:!MD5; ssl_prefer_server_ciphers on; # Pass the PHP scripts to FastCGI server listening on 127.0.0.1:9100 location ~ \.php$ { try_files $uri =404; fastcgi_param HTTPS on; fastcgi_pass 127.0.0.1:9100; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include fastcgi_params; } Contents of location.conf: # Remove trailing slash to please Laravel routing system. if (!-d $request_filename) { rewrite ^/(.+)/$ /$1 permanent; } location / { try_files $uri $uri/ /index.php?$query_string; } # We don't need .ht files with nginx. location ~ /(\.ht|\.git|\.svn) { deny all; } # Added cache headers for images. location ~* \.(png|jpg|jpeg|gif)$ { expires 30d; log_not_found off; } # Only 3 hours on CSS/JS to allow me to roll out fixes during early weeks. location ~* \.(js|css)$ { expires 3h; log_not_found off; } # Add expire headers. location ~* ^.+.(gif|ico|jpg|jpeg|png|flv|swf|pdf|mp3|mp4|xml|txt)$ { expires 30d; } # Pass the PHP scripts to FastCGI server listening on 127.0.0.1:9100 location ~ \.php$ { try_files $uri /index.php =404; fastcgi_index index.php; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; include fastcgi_params; fastcgi_pass 127.0.0.1:9100; } Any ideas where this is going wrong?

    Read the article

< Previous Page | 359 360 361 362 363 364 365 366 367 368 369 370  | Next Page >