Search Results

Search found 20935 results on 838 pages for 'content'.

Page 50/838 | < Previous Page | 46 47 48 49 50 51 52 53 54 55 56 57  | Next Page >

  • how to set title within content page asp.net

    - by varshney4u
    Hi, I have created a Content Page using Master Page. Within Master Page, I have created the following tag for Title: <head id="Head1" runat="server"> <title></title> <meta http-equiv="X-UA-Compatible" content="IE=EmulateIE7" /> </head> Within Content Page, following are set: protected void Page_Load(object sender, EventArgs e) { if (!Page.IsPostBack) { if (String.IsNullOrEmpty(Request.QueryString["PatientRegistrationKey"])) { // .... } else { this.Page.Title = "My New Title"; //.... } } Though I am also setting the Master Page at run time as bellow: protected void Page_PreInit(Object sender, EventArgs e) { if (String.IsNullOrEmpty(Request.QueryString["PatientRegistrationKey"])) this.MasterPageFile = "~/MasterPages/A.master"; else this.MasterPageFile = "~/MasterPages/B.master"; } While getting open this page in browser, I got following Title: http://localhost:3562/?PatientRegistrationKey=0 - My New Title Please advice for the changes, so that there should be only My New Title within title, nothing extra like query string etc. Any help would be appreciated.

    Read the article

  • Vertically aligning CSS :before and :after content

    - by danixd
    I am trying to centre the link with the image, but can't seem to move the content vertically in any way. <h4>More Information</h4> <a href="#" class="pdf">File Name</a> The icon is 22 x 22px .pdf{ } .pdf:before{padding:0 5px 0 0; content: url(../img/icon/pdf_small.png);} .pdf:after{content: " ( .pdf )"; font-size: 10px;} .pdf:hover:after{color: #000;}

    Read the article

  • How to enable 'wget' to download the whole content of HTML with Javascript

    - by neversaint
    I have a site which I want to download using Unix wget. If you look at the source code and content of the file it contain section called SUMMARY. However after issuing a wget command like this: wget -O downdloadedtext.txt http://www.ncbi.nlm.nih.gov/IEB/Research/Acembly/av.cgi?db=mouse&c=gene&a=fiche&l=2610008E11Rik The content of the downdloadedtext.txt is incomplete and different with the source code of that site. For example it doesn't contain SUMMARY section. Is there a correct way to obtain the full content correctly? The reason I ask this because I want to automate the download from different values in that HTML.

    Read the article

  • Notify a content resolver when normal sql insert or update

    - by user1400538
    I have created a small content provider (which I only use to query a db table and fetch some value). I have confirmed that the provider works fine. The table values get updated on a regular basis through normal sql insertions(and not through anycontent provider) Whenever an insert/update or delete occurs through a normal sqlite operation as mentioned above, I need to notify the content resolvers which was written to communicate with the content provider just for quering database and fetch some values. Is this, possible? If yes what need to be done? Any help is appreciated.

    Read the article

  • Android Read contents of a URL (content missing after in result)

    - by josnidhin
    I have the following code that reads the content of a url public static String DownloadText(String url){ StringBuffer result = new StringBuffer(); try{ URL jsonUrl = new URL(url); InputStreamReader isr = new InputStreamReader(jsonUrl.openStream()); BufferedReader in = new BufferedReader(isr); String inputLine; while ((inputLine = in.readLine()) != null){ result.append(inputLine); } }catch(Exception ex){ result = new StringBuffer("TIMEOUT"); Log.e(Util.AppName, ex.toString()); } in.close(); isr.close(); return result.toString(); } The problem is I am missing content after 4065 characters in the result returned. Can someone help me solve this problem. Note: The url I am trying to read contains a json response so everything is in one line I think thats why I am having some content missing.

    Read the article

  • django - change content inside an iframe

    - by fabriciols
    Hello guys, I have a site running in django and want to make it available on third party sites. In these sites they make my content available through an tag. When my site is requested from within that iframe I would like to modify the content (remove header, filter data, etc.). What would be the best way to do this? Is there any way of knowing that the request is being made from an iframe? Multiples sites will request the same url, can I change the content, based on the requesting site? Thanks! PS: Sorry for my bad english :/

    Read the article

  • druapal content image not showing

    - by tibin mathew
    hi, I'm doing a drupal site. i want show a facebook image in my page content in home page. i have edited content of that page. i have edited the body of that content through admin and put the text in body like this The BikeGlow™ Safety Light runs on 2-AA batteries. Light tube length is 10'. Complete unit weighs under 2 oz (without batteries). Light tube attaches to frame with included attachment strips. "It's like magic cycling protection from cars!" < img src='bikeglow_com/sites/default/files/5u84f48n.gif' alt="facebook" / but after saving image is not showing there even that alt tag is not coming. can any one help me, what will be reason for his

    Read the article

  • drupal content image not showing

    - by tibin mathew
    hi, I'm doing a drupal site. i want show a facebook image in my page content in home page. i have edited content of that page. i have edited the body of that content through admin and put the text in body like this The BikeGlow™ Safety Light runs on 2-AA batteries. Light tube length is 10'. Complete unit weighs under 2 oz (without batteries). Light tube attaches to frame with included attachment strips. "It's like magic cycling protection from cars!" < img src='bikeglow_com/sites/default/files/5u84f48n.gif' alt="facebook" / but after saving image is not showing there even that alt tag is not coming. can any one help me, what will be reason for his

    Read the article

  • Slide div content from left to right

    - by Inuka Rathnayake
    I'm beginner in to jquery and i wonder always how do developers do like this. Can someone give me a simple example..? http://static.livedemo00.template-help.com/wt_37928/#!/home What i need is there is a menubar and when you click a link from it that link's content should come to center of the screen by left and when you click another link that visible content should go to right and recently clicked link's content come to the screen as i mentioned. Can anyone help me out...? It will be a great relieve to me. Thank you.

    Read the article

  • Indexing an image in-between already made content

    - by Christophersson
    I have a pre-code page coded as follows: <div id="linearBg"> <div id="wrapper"> <div class="logo"></div> <div class="navigation"></div> <div class="video"></div> <div class="content"></div> </div> </div> Where linearBg is a gradient background, the back board of the website. Wrapper is the container for the inner div's, and the rest are content oriented. So i've already implemented this with styles and all sorts, but the thing is I want to add: <div class="watermark"></div> underneath/behind both the content and video div, sort of like a reverse watermark, I've tried z-indexing but i'm not an expert. Could you guide me on to do make this possible? Thanks in advance

    Read the article

  • Website Content via remote mysql or API for clients

    - by Kannan
    I would like to know whether there will be problem for shared hosting/vps/dedicated users if they request content via remote mysql or via http API. ? What will be issues if they make every request to production server for delivery of the website articles along with image urls etc. Will it cause disk wait on shared/vps/dedicated hosting users ? Will both remotemysql or http api or curl cause high CPU usage, I/O wait or Memory usage ? What should be best for speed remote mysql or http web api ? or any alternatives ? Thanku Kannan

    Read the article

  • IIS reveals internal IP address in content-location field - fix

    - by saille
    Referring: http://support.microsoft.com/kb/q218180/, there is a known issue in IIS4/5/6 whereby it will reveal the internal IP of a web server in the content-location field of the HTTP header. We have IIS 6. I have tried the fix suggested, but it has not worked. The website is configured to send all requests to ASP.NET, and I am wondering if this is why the fix, which addresses IIS configuration, has not worked for us. If this is the case, how would we fix this in ASP.NET? We need to fix this issue in order to pass a security audit.

    Read the article

  • IIS reveals internal IP address in content-location field

    - by saille
    Referring: http://support.microsoft.com/kb/q218180/, there is a known issue in IIS4/5/6 whereby it will reveal the internal IP of a web server in the content-location field of the HTTP header. We have IIS 6. I have tried the fix suggested, but it has not worked. The website is configured to send all requests to ASP.NET, and I am wondering if this is why the fix, which addresses IIS configuration, has not worked for us. If this is the case, how would we fix this in ASP.NET? We need to fix this issue in order to pass a security audit.

    Read the article

  • mod_cache serving the wrong content

    - by J. Pablo Fernández
    I'm trying to use mod_disk_cache to speed up a web site that is running on WordPress. Whenever I enable it with CacheEnable disk / and the rest being the stock Ubuntu configuration I start to get the wrong results. When I see the main page it's fine, but when I go to a specific post I get a RSS feed instead. Like if the cache is returning the wrong content. I've disabled my RewriteRules as it seems mod_cache doesn't work with that. I'm not even sure where to start to debug such a thing. Any ideas?

    Read the article

  • ASP.NET MVC app on IIS7 with WebForms content is throwing NTLM authenticate box

    - by Jon
    I have an ASP.NET MVC app that also contains some WebForms content (for SSRS ReportViewer). This is deployed to IIS7 and the MVC pages of the app work fine, but when I try to browse to the aspx page I am prompted with the NTLM auth box. I do not have NTLM enabled, I only have Anonymous auth enabled. I have this deployed and fully working on an IIS6 box, the only other difference is that the IIS6 box is in our company domain, but the IIS7 box is not (I fail to see how this could be the issue as the MVC stuff is working fine). Any thoughts? Thanks.

    Read the article

  • Windows 7 x64 Pro not indexing PDF content ?

    - by Wookai
    In an attemp to get rid of all the paper clutter that I receive each day, I started scanning and ORCing all my important documents last year. I was using Windows Desktop Search on Vista to index my "Administrative" folder in which I keep all these scanned documents, to be able to quickly retreive them when needed. I upgraded to Windows 7 RTM recently (I got the x64 Pro version, via the MSDN-AA channel of my university). Since then, it seems that the content of the PDFs is not indexed anymore. For example, I have the warranty for my GTX260 graphic card. In the PDF, using Acrobat Reader, I can search for the string "GTX260" and it finds the relevant lines. However, if I perform the same search in the folder containing the PDF, nothing is found. I checked in the indexing options, the folder is supposed to be indexed.

    Read the article

  • How to sum cells depending on the content of a neighbor cell

    - by dannymcc
    I have an Excel document with the following columns; Date | Reference | Amount 23/01/11 | 111111111 | £20.00 25/09/11 | 222222222 | £30.00 11/11/11 | 111111111 | £40.00 01/04/11 | 333333333 | £10.00 31/03/11 | 333333333 | £33.00 20/03/11 | 111111111 | £667.00 21/11/11 | 222222222 | £564.00 I am trying to find a way of summarising the content in the following way; Reference : 111111111 Total: £727 So far the only way I have been able to achieve this is to filter the list by each reference number (manually) and then add a simple SUM formula to the bottom of the list of amounts. Are there any tricks that anyone knows that may speed this up? What I am trying to achieve is a spreadsheet that highlights each reference number that collectively exceeds over £2,000.

    Read the article

  • Windows 8, Truecrypt drive mounts but Windows cannot read its content

    - by phil
    I installed Windows 8 and when I started it, it said that it was repairing a disk. The disk was an encrypted Truecrypt disk. I couldn't mount the disk in Truecrypt after that. I tried to repair the header and it worked, now I can mount the disk but neither Windows 7 or 8 can read the content. Windows asks if it should format the disk. I have all the important files on backup, but there are some media files that I would like to get back. Any ideas?

    Read the article

  • View unknown content-types in Firefox?

    - by dirtside
    A certain URL on my server returns Content-Type: application/json. The filename ends with .phtml, so whenever I go to it, Firefox 3.5 asks me if I want to save it or open it with another program. But the answer is neither; I want to view it in Firefox as if it was text/plain. Suggestions? The Applications tab of the Preferences window doesn't give this as an option for the "PHTML" type which now is listed in there (ever since I tried to open it with no external program).

    Read the article

  • Nginx Cache-Control

    - by optixx
    Iam serving my static content with ngnix. location /static { alias /opt/static/blog/; access_log off; etags on; etag_hash on; etag_hash_method md5; expires 1d; add_header Pragma "public"; add_header Cache-Control "public, must-revalidate, proxy-revalidate"; } The resulting header looks like this: Cache-Control:public, must-revalidate, proxy-revalidate Cache-Control:max-age=86400 Connection:close Content-Encoding:gzip Content-Type:application/x-javascript; charset=utf-8 Date:Tue, 11 Sep 2012 08:39:05 GMT Etag:e2266fb151337fc1996218fafcf3bcee Expires:Wed, 12 Sep 2012 08:39:05 GMT Last-Modified:Tue, 11 Sep 2012 06:22:41 GMT Pragma:public Server:nginx/1.2.2 Transfer-Encoding:chunked Vary:Accept-Encoding Why is nginx sending 2 Cache-Control entries, could this be a problem for the clients?

    Read the article

  • How do i mitigate DDOS attacks on static servers?

    - by acidzombie24
    Here is a slightly different take on DDOS attacks. Rather than a server with dynamic content being attack i was curious how to deal with attacks on servers with STATIC CONTENT. This means cpu tends to not be an issue. Its either bandwidth or connection problems. How would i mitigate a DDOS attack knowing nothing about the attacker (for example country, ip address or anything else). I was wondering is shorting the timeout and increasing amount of connections is an acceptable solution? Or maybe that is completely useless? Also i would limit the amount of connections per IP address. Would the above help or be pointless? Keeping in mind everything is static checking for multiple request of the same page (html, css, js, etc) could be a sign of a attack. What are some measures i can take on a static content server?

    Read the article

  • Adobe Reader - Content Preparation progress

    - by kubal5003
    Hello, I was wondering recently if is it just me or anyone else noticed it: why Adobe Reader after opening every single document displays "Content preparation" window with progress bar and it lasts for ages.. ? On linux pdf readers work hell lot better (faster), on windows other readers also work faster. Some years back in the past Adobe Reader also used to be quick. What has happened? PDF files aren't bigger/more complex compared to 3-4 years ago. Computers are at least dual core and with much more ram and displaying pdf files is getting slower and slower..

    Read the article

  • Fastest web server for serving static content

    - by Swader
    Hello, I'm optimizing our system for some faster static content delivery, and was wondering if anyone has any proper experience with the fastest web servers out there for such a purpose. From the three main candidates I've considered, Nginx, Cherokee and Lighttpd, each seems to have its own problems - but the reports I've read online are somewhat biased and lean towards whichever server the user is currently using. Any ideas on where to see a proper benchmark for this specific purpose, or at least a non-biased list of pros and cons? Any personal experiences and pitfalls I should be vary of? Thanks Edit: Serverfault.com gave the answer as nginx. I'd still like to hear some developer thoughts from this end of the universe.

    Read the article

  • How to disable Spotlight content indexing in Mac OS

    - by o.v.
    From Windows experience, I could always elect Live search to only index file names not their content. Is this something that can be done with Spotlight on a Mac? It used to index absolutely everything, for instance it would return a bunch of video files for any obscure character combination typed into the search field. Right now I've disabled Spotlight entirely as per this answer, but it seems to have disabled searching altogether. For instance, Finder is yet to locate any .pdf files in a small directory as I'm typing this question (unlike windows search which would still be able to work even with indexing disabled) Alternatively, if there is any way (including a trusted third-party app) that will index file names and metadata e.g. ID3 tags that would likely be the preferred option.

    Read the article

  • Can Squid 2.7 proxy gzipped content

    - by Tom Styles
    We have a forward proxy for our network which is Squid 2.7. This is managed for us by a third party. We noticed recently that http requests going from our network to the web were having the Accept-Encoding header removed. This was resulting in all web traffic across our network (approx 8000+ PCs) being uncompressed even though the browsers and server on each end were capable. We have asked the third party to look into this and they have said it is because Squid 2.7 does not support compression. I understand this to be true but I was under the impression that the compression happened on the webserver rather than the proxy. So... Can Squid 2.7 proxy and/or cache content that is gzipped? If it can, how/why might it be configured such that the Accept-Encoding header is being removed?

    Read the article

< Previous Page | 46 47 48 49 50 51 52 53 54 55 56 57  | Next Page >