Search Results

Search found 29495 results on 1180 pages for 'cross site scripting'.

Page 209/1180 | < Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >

  • Alternate User select interface in django admin to reduce page size on large site?

    - by David Eyk
    I have a Django-based site with roughly 300,000 User objects. Admin pages for objects with a ForeignKey field to User take a very long time to load as the resulting form is about 6MB in size. Of course, the resulting dropdown isn't particularly useful, either. Are there any off-the-shelf replacements for handling this case? I've been googling for a snippet or a blog entry, but haven't found anything yet. I'd like to have a smaller download size and a more usable interface.

    Read the article

  • different small js files per page VS. 1x site-wide js file?

    - by Haroldo
    Different pages of my site have different js needs (plugins mainly), some need a lightbox, some dont, some need a carousel, some dont etc. With regards to pageloading speed should i option 1 - reference each js file when it is needed: so one page might have: <script type="text/javascript" src="js/carousel/scrollable.js"></script> <script type="text/javascript" src="js/jquery.easydrag.js"></script> <script type="text/javascript" src="js/colorbox/jquery.colorbox-min.js"></script> and another have: <script type="text/javascript" src="st_wd_assets/js/carousel/scrollable.js"></script> <script type="text/javascript" src="st_wd_assets/js/typewatch.js"></script> option 2 - combine and compress into one site_wide.js file: so each page would reference: <script type="text/javascript" src="js/site_wide.js"></script> there would be unused selectors/event listeners though, how bad is this? I would include any plugin notes/accreditations at the top of the site_wide.js file

    Read the article

  • Subversion: setting up a remote repository and running my site off it?

    - by Matt Andrews
    Hi all. I'm new to SVN and have experimented with it locally on my Dreamhost test server (which has a Subversion "one-click-install" function). Having found my way around the functionality I'm definitely sold, but a little lost about using it to manage my work website (not hosted with Dreamhost, so not offering a one-click SVN installation). Am I correct in thinking that I can set up a repository on my website root (which contains all the files), and then when I develop new features and run a commit, this will update my site? Is this the proper workflow for this sort of thing? If so, is there a standard way to set this kind of thing up on my remote server? Thanks.

    Read the article

  • How can I quickly set up a RESTful site using PHP without a Framework?

    - by Sanoj
    I would like to quickly set up a RESTful site using PHP without learning a PHP Framework. I would like to use a single .htaccess-file in Apache (no mod_rewrite) or a single rule using Nginx, so I easyli can change web server without changing my code. So I want to direct all requests to a single PHP-file, and that file takes care of the RESTful-handling and call the right PHP-file. In example: The user request http://mysite.com/test The server sends all requests to rest.php The rest.php call test.php (maybe with a querystring). If this can be done, is there a free PHP-script that works like my rest.php? or how can I do this PHP-script?

    Read the article

  • How to make one ASP.NET MVC site "derive" from another.

    - by Rob Levine
    My question is similar to "ASP.NET 2 projects to share same files", but with an ASP.NET MVC slant. Basically, we have two sites, one being based mostly on the other (roughly 90% views, controllers, images, in the second are identical to the first). However, in some cases, the views may be different, or a controller in the second site may be different to the first. Are there any simple ways of achieving this in ASP.NET MVC? So far, we've looked at using linked files to have two totally seperate projects where the second project shares the files it needs from the first. One problem with this approach is that most pages in the second project don't literally exist in the virtual directory, it makes debugging a pain - you have to publish in order to generate the files so you can debug. Does anyone have any better approaches, or ways this approach can be simplified?

    Read the article

  • WordPress page is listed in "Posts" but appears on site as a "Page", What gives?

    - by Scott B
    I've seen this on a few sites now and its puzzling me. The page will be listed under the "Edit Posts" listing, however, when I click on "View Page" to see the live post, its being served out as a "Page". I can tell this by viewing the source and the body tag is... <body class="page page-id-37 logged-in"> If its were actually a post, the body tag should be <body class="single postid-63 logged-in"> I've seen a "post to page" and a "page to post" plugin that will convert between pages and posts, however, this site has NO plugins other than the wordpress defaults (akismet and hello dolly). I'm stumped.

    Read the article

  • Has Google introduced a system that allows multiple "keyword-only" domains per site?

    - by tags2k
    I've been told by a client that "a friend" told them that as of January 2010, Google allows multiple domain names that have keywords in them to be associated with a single site. To be honest this sounds rather April-foolish but I'm not sure when his "friend" told him so for the time being I have to take it at face value. I've heard nothing of this and have searched for such a thing this morning, to find nothing but warnings against this practice. Said client seems keen on buying up lots of domains today, so before he insists upon it I just want to be absolutely sure - has Google silently introduced such an allowance, or is there something else they introduced earlier this year that he could be getting confused with? Thanks for any light you can shine on this!

    Read the article

  • How to get response from secure site (https) using curl in php?

    - by Chirag
    Hi, i have used below code for reading the response from HTTTPS secured site using curl but it shows an error like 'Page can't be displayed" curl_setopt($ch, CURLOPT_HTTPAUTH, CURLAUTH_ANY); curl_setopt($ch, CURLOPT_SSL_VERIFYPEER, true); curl_setopt($ch, CURLOPT_SSL_VERIFYHOST, true); curl_setopt($ch, CURLOPT_URL, $gatewayURI); curl_setopt($ch, CURLOPT_HEADER, 1); curl_setopt($ch, CURLOPT_RETURNTRANSFER, 1); curl_setopt ($ch, CURLOPT_POST, 1); curl_setopt ($curl_ch, CURLOPT_CAINFO, dirname(__FILE__)."/tempcert.pem"); any help??

    Read the article

  • Facebook page linking to external site sign-up process, capture permission to write to wall in process?

    - by steve
    Hi all, Have had a good hunt through the archive but can't find anyone trying to do this... hope someone familiar with the facebook API can confirm if it's possible? Basically I have a client who wants to replicate their membership sign up process in a tab on their facebook page. The form would still submit to their own website to process, we'd just be replicating the form fields. As an additional requirement they want to capture peoples facebook user ID and get permission to post back to a users wall at the same time... The idea being that once the user is a member we can post back to their wall so their friends see that they've signed up... Basically after a sanity check that: 1) these things are possible to do; 2) the best method to build the form in a FB page - I'm guessing using JS to create all fields & ajax to submit to the external site? Thanks Steve

    Read the article

  • A fast way to get Reporting Services local site working?

    - by Junior Mayhé
    Hello I was here trying to figure out why my Reports manager is empty, there's no tabs at all. I installed SQL Server 2008 complete, but didn't not configure Reporting Services. When installing SQL Server 2008, this Windows 7 version didn't have yet IIS installed, I installed it later. I don't see where is this localhost/Reports physically on my Hard Drive, where is the physic folder? I don't see on IIS where is Report folder, would it exist? The site settings people talk about, I can't find it. The "Reporting Services" service is running on automatic at SQL Server Configuration Manager. How can I get Reporting Services this working without struggling? (I can't see why is necessary to customize all these user permissions)

    Read the article

  • SEO: It's recommended to upload and put live a beta / non-finished version of web site??

    - by Jonathan
    I'm working on this big website and I want to put it online before its fully finished... I'm working locally and the database is getting really big so I wanted to upload the website and continue to work on it in the server, but allowing people to enter, so I can test. The question is if this is good for SEO, I mean, there are a lot of things SEO related that are incomplete.. For example: there are no friendly URLs, no sitemap, no .htacces file, lot of 'in-construction' sections... Does Google will penalize me forever? How does it works? Google indexes adn get the structure of the site just once or its constantly updating and checking for changes??? What i should do? What you recommend?!!?

    Read the article

  • What makes a web site 'finished' for delivery to a client?

    - by AP257
    Hi, Sorry if this question has already been answered, and sorry if it's too subjective to make sense, or for StackOverflow. I'm delivering a site to a client and I want to send them high-quality HTML/CSS/JS, fully validated, accessible etc. So I'm compiling a list of things to check, and useful tools for doing so, before I hand the code over. Here's a partial list, but what am I missing? Link checking for any broken links I might have missed - W3C link checker HTML validation for accessibility and broken-ness - W3C HTML validator CSS validation - W3C CSS validator Check for slow-loading page elements - Firebug and YSlow plugin What's missing - if you were a client, what else would you want to be sure has been checked? I'm wondering about the etiquette of things like comments, indentation, and minification; is it good practice to sort out all these? And what else have I missed? Thanks :)

    Read the article

  • What is the correct HTTP status code to send when a site is down for maintenance?

    - by alex
    Is there a HTTP status code to tell Google (and others) to go away, index me again later? Basically, one that semantically tells clients that the site is down for maintenance? The ones I have considered are 304 => Not modified 307 => Temporary redirect 410 => Gone 503 => Service Unavailable I'm leaning towards the last one, but was just curious as to which one was proper choice. Thanks Update Is this the correct way to send it with PHP? header('Status: 503 Service Unavailable');

    Read the article

  • Using Wordpress as a CMS to a site AND blog (two completely different layouts for each). Duplicate i

    - by Somebody still uses you MS-DOS
    I'm going to create a website using Wordpress static pages feature. It's going to have a menu for all the created pages. A blog is going to be created as well. This blog has a completely different layout/menu from the main site, but within the same domain: "http://www.domain.com" is the website url, and "http://www.domain.com/blog/" is the blog url. I was thinking of installing two different Wordpress instances: /var/www/public_html/ and /var/www/public_html/blog/. Although it's a simple solution, users are going to be duplicated: if you change password in one system, you need to change in the other. So, I would like to know from you which other approaches are possible: create a "dynamic" theme (call a theme when in root domain, another whern /blog - dont even know if this is possible), or for example a plugin to syncronize users from both instances? Does Wordpress MU solve this problem for me? Thanks!

    Read the article

  • PHP error logging - can I log all ofg the site's error to a single file?

    - by mawg
    hi, in PHP.INI, I can set the variable "error_log" and it is currently set to "error_log", which manes one file of that name in each directory Do you know ... if I set it to "public_html/error_log" - will I get only one single error log file? Any other way to do so? I really just want a single site-wide error file to check, rather than one on each sub-directory. For bonus marks, can I send myself an email each time a new entry is added to the file? Left as an exercise for the reader - can I ignore some "errors" which aren't really?

    Read the article

  • How can I make a boring project (another WordPress site) interesting?

    - by Christopher Altman
    WordPress is my example, but the question can be generalized to any technology that is not particularly interesting. To me, WordPress takes away the intellectually gratifying pieces of coding. I would rather write a new version of WordPress than write a WordPress theme and glue together some plugins. I am using WP because my company dictates the platform for some of our clients (I do not disagree with the choice from a business perspective, WP is quick and cheap to implement). My question is, how can I make my next WordPress project interesting? I want to advance my understanding of the fundamentals of programming (aka data structures, algorithms, and caching) but do not see how I can achieve this when coding another WP site. I have a fairly tight understanding of front-end technologies and believe I have made WP do things it was never intended to do. Examples are here and here. Solving front-end related problems is not as interesting as coding a full stack application. Any advice will help.

    Read the article

  • Site works perfect in Mozilla but not in IE. Is my js file not compatible with IE

    - by Bonkers
    I'm working on a site written in PHP/MySQL. We have a form to reserve time on a calendar and it works great in Mozilla and stores the reservation to our database, but in IE you fill out the form and when you click the "Reserve" button to submit it and nothing happens. All I can think of is that my javascript is not working with IE. I have these lines in my .js file: resLenT = document.getElementById(resLenElem); resLenI = resLenT.selectedIndex; resLen = resLenI + 1; where resLenElem is a drop-down box. These are the only lines that I can think of at the moment that might be causing trouble in IE. Does this all sound like I'm on the right track or am I way off base?

    Read the article

  • How can i add Active Directory security groups to a SharePoint site to control permissions, rather than individual user accounts

    - by user574811
    SharePoint does integrate active directory accounts, of course, but how about security groups? Have a few sites where I'm fairly confident access is going through an existing Active Directory (AD) security groups (i.e. only an AD security group has been granted permissions through the 'People and Groups') In another situation, where I created the AD group and granted it permissions to a site, the customers were not able to access immediately. Eventually had to fast-track it and add the individuals to the People and Groups to keep the project going, but hoping not to have to maintain it that way. Any specific requirements of the security group in AD? Universal, Global, or domain local? Is there any time delay between modifying group members in AD and having that take effect in SharePoint?

    Read the article

  • Optimal ASP.Net cache duration for a large site?

    - by HeroicLife
    I've read lots of material on how to do ASP.Net caching but little on the optimal duration that pages should be cached for. Let's say that I have a popular site with 50,000 pages. The content does not change frequently, so I could cache pages for up to an hour if I wanted. The server has 16 GB of RAM, but database connections are limited. How long should pages be cached for? My thinking is that if I set the cache duration too high (let's say 60 minutes), I will fill up memory with a fraction of the total content, which will continually be shuffled in and out of memory. Furthermore, let's say that 10% of the pages are responsible for 90% of traffic. If the popular pages are hit every second, and the unpopular ones every hour, then a 60 second cache would only keep the load-intensive content cached without sacrificing freshness. Should numerous but rarely-accessed content be cached at all?

    Read the article

  • Windows Ce 6.0 loses Windows credentials when viewing a web site that's running on Windows 2008 server

    - by gnomixa
    When a user views a web page (with integrated Windows authentication) on WindowsCE 6.0 device, the authentication is lost sporadically. The page being viewed is running on Windows 2008 server. We never had the same issue with Windows 2003 server. The credentials were being asked once and cached for a certain time. My question is: has anything changed in Windows 2008 that doesn't pass the credentials the same way to WindowsCE? The only variable in this scenario is the web server OS - Windows 2003 vs WIndows 2008. Any help would be appreciated, thanks!

    Read the article

  • Setup LAN to serve webpages and voip and access to the web site from inside LAN with domain name

    - by Mauricio Arias
    I'd like to know if it will work: I have my domain and I´m serving a webpage in a nginx to the internet, but if I type my domain in my laptop inside LAN I access to my modem/router configuration, I cannot access to the web server unless I type the IP address. I would like to add a Bind server after the modem/router - (port forward, ports 80 and 5060), if the request is www.mydomain.com bind should resolve the nginx IP address and serve it, and if it is a voip request should address to the voip server and if I'd like to access to the website from inside LAN I'd like to type mydomain.com. Could I do it with this configuration? Do I need something else? Thanks in advace!

    Read the article

  • Is there a Firefox or Chrome plugin, or a standalone program, for monitoring site usage and search queries?

    - by Leigh Caldwell
    I'm running some research on how people search the web for specific types of information. I'd like to be able to set them up with a laptop and browser and then record a history of what they search for and what sites they visit. A Firefox or Chrome plugin would be ideal, but a standalone program is fine too. It doesn't need to be free, just quick and reliable. It doesn't need to be a general PC monitoring program (though that would be OK too) - it's only Web usage I need to track. I've found a few on the Web but am not sure which ones to trust. Your recommendations would be much appreciated.

    Read the article

  • Using dnsmasq for accessing multiple nameservers assigned by DHCP

    - by Ash
    At my work desktop running openSUSE 11.4, I have a local network which gets its address, domain (work.site) and nameservers (10.100.1.1, 10.100.1.2) info through DHCP - which get written into /etc/resolv.conf I get to access the internet using the work network, and these 2 nameservers end up returning the entries for any public domain name lookups on the internet. I also have a private VPN that I end up connecting. The nameserver (10.111.1.1) and domain (private.site) are rarely bound to change for this network, but currently they're pushed by the openVPN client into networkmanager, and which also gets merged with the existing /etc/resolv.conf My resolv.conf ultimately ends up looking like this: search private.site work.site nameserver 127.0.0.1 nameserver 10.111.1.1 nameserver 10.100.1.1 As you can see the 2nd nameserver from my work network was pushed out because of the max 3 entry limitations. It is fine still, but would be a problem if that nameserver goes down for maintenance or something. So I found out that dnsmasq could help me here, and hence I setup dnsmasq just as a local DNS resolver without any DHCP support. So right now this is my /etc/dnsmasq.conf: resolv-file=/etc/resolv.conf server=/private.site/10.111.1.1 server=/1.111.10.in-addr.arpa/10.111.1.1 listen-address=127.0.0.1 bind-interfaces log-queries I've made dnsmasq get the list of nameservers from /etc/resolv.conf since NetworkManager seems to be updating this list correctly (for a max of 3 nameservers). I'm able to resolve the host names in both the networks correctly. So these are the questions I have: Is there a way I can make either NetworkManager or dhclient write out the list of nameservers somewhere else which I can make dnsmasq use as resolv-file ? How do I make dnsmasq use certain nameservers as the default for all queries ? Right now I notice that lookups for public domains on the internet are usually sent to both the nameservers - the one on work.site as well as private.site. It would be good if I can limit this only to work.site.

    Read the article

  • Having XP VM use my host OSX ssh tunnel to connect to a remote site?

    - by Manachi
    I am using Mac OSX and have Windows XP running on VMWare Fusion. I'm creating an ssh tunnel from OSX to a remote server, and then trying to have Windows XP use that tunnel (I actually use a program called Proxifier on XP to filter my XP MS SQL Server traffic through that tunnel) Note that I can successfully create an ssh tunnel (on port 9333) from the XP putty to the remote host, and have SQL Server Proxify through that tunnel and it all works correctly. However when I try to set up the tunnel in OSX, and have Proxifier in XP point to the OSX tunnel instead of localhost, it doesn't seem to connect. Here is the OSX command i'm using to create the tunnel: ssh -i /my/key -p 9001 -D 9333 -g me@remotehostname Then I set my XP proxifier to point to macosxhostname:9333 (instead of the previous localhost:9333 which worked corrently when using putty) Any suggestions on what I may have missed? My XP firewall is turned off while setting this up.

    Read the article

  • How can I edit an individual site's virtualhost using either WHM or cPanel?

    - by user55578
    I've been using Webmin/virtualmin for years. In Webmin, I can edit Apache config files quite easily. For example, if a user wants to change the DocumentRoot because he/she wants to serve up a Ruby on Rails app using Phusion Passenger, I can do that in a few seconds using the Webmin GUI. /etc/apache/sites-available/samplesite.com.conf Is there something similar in WHM/cPanel? How can I edit the VirtualHost (and inside that, the Document Root), using WHM/cPanel?

    Read the article

< Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >