Search Results

Search found 45340 results on 1814 pages for 'page title'.

Page 118/1814 | < Previous Page | 114 115 116 117 118 119 120 121 122 123 124 125  | Next Page >

  • Why Does Thread.CurrentThread.CurrentCulture Change between Page Rendering and HttpModule.PostReques

    - by Chad
    I'm creating an HttpModule that needs to know the value of Thread.CurrentThread.CurrentCulture as set in an MVC application. That value is currently being set by the BaseController, but when my HttpModule.PostRequestHandlerExecute() method fires, it reverts to what the Culture was prior to page rendering. I have duplicated this by creating a simple web app with these steps: Module.PreRequestHandlerExecute: Set culture to A Page_Load: Culture is currently A. Set culture to B Module.PostRequestHandlerExecute: Current thread culture is A. I expected it to be B but it was changed between page rendering and PostRequestHandlerExecute Any idea why .Net changes this value or how I could get around it? The thread is the same, so something in .Net must be explicitly reverting the culture.

    Read the article

  • Window Media Player issues two requests for the audio on web page

    - by Ron Harlev
    I'm using Windows Media Player in a web page. I have version 11 installed so that is the version I'm testing with right now. The player is embedded on the page with this HTML: <OBJECT id='MS_mediaPlayer' width="400" height="45" classid='CLSID:6BF52A52-394A-11D3-B153-00C04F79FAA6' codebase='http://activex.microsoft.com/activex/controls/mplayer/en/nsmp2inf.cab#Version=5,1,52,701' standby='Loading Microsoft Windows Media Player components...' type='application/x-oleobject'> <param name='autoStart' value="false"> <param name='uiMode' value="invisible"> <param name='loop' value="false"> </OBJECT> I'm calling in JavaScript: MS_mediaPlayer.URL = "SomeAudioFile.mp3" MS_mediaPlayer.controls.play(); When I look at Fiddler I can see that the player actually downloads "SomeAudioFile.mp3" twice. Is there some setting I have wrong? I was trying to set the "autoPlay" to true and avoid calling "play()". Got the same result - two downloads. UPDATE: The first request's user-agent is "Windows-Media-Player/11.0.5721.5268". The second has "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; GTB6; .NET CLR 1.1.4322; .NET CLR 2.0.50727; .NET CLR 3.0.04506.30; .NET CLR 3.0.04506.648; .NET CLR 3.5.21022; .NET CLR 3.0.4506.2152; .NET CLR 3.5.30729)". Looks like the browser is running the same request the second time. No Idea why Any ideas? UPDATE (4/1/10): Still no solution. I debugged the JS thoroughly and there is only one call to MediaPlayer.URL='.....' to set the audio file. Nothing else triggers the media player to load the file and there is no other place referencing the audio file on the page. One other interesting fact is that this doesn't happen (the double loading of the audio) when I run the browser locally on my development web server. But other remote requests to the same web server generate the double audio loading. I believe I eliminated any correlation with specific IE version or media player version. This happens with IE6-8 and WM9-12

    Read the article

  • How to customize the Visual Studio 2010 RC StartPage recent projects

    - by devlife
    I'd like to play around with customizing the visual studio 2010 rc start page recent items. For what I have in mind I'd need to customize the datasource / databinding but I can't find where the information is coming from. <ScrollViewer Grid.Row="1" HorizontalAlignment="Stretch" Style="{DynamicResource StartPage.ScrollViewerStyle}" VerticalAlignment="Stretch" VerticalScrollBarVisibility="Auto"> <sp:MruListBox DataContext="{Binding Path=RecentProjects}" ItemsSource="{Binding Path=Items}" Background="Transparent" BorderThickness="0" AutomationProperties.AutomationId="MruList"/> </ScrollViewer> Can anyone point me in the right direction? I see that it is binding to RecentProjects but where is that coming from?

    Read the article

  • Client side page permissions - Javascript / jquery / cookies / other?

    - by Ozaki
    TLDR Using plain HTML / Javascript. Want to block access to some pages (doesn't have to be super secure just to stop some peeking eyes). I thought of simply doing this by setting a cookie for each page they are allowed to visit with a value of true but thats a bit messy. Although it would work. Is there a way to set an array of values to a cookie so I can read the cookie and if a name of a page is in there then allow access with an IF statement or so on each of my pages. If they dont have the cookie just to replace my #content (entirepage) to "sorry no" etc. For example: $.cookie("Access","page1, page2, page3",{ expires: 1 }); Am already using JQuery, Jquery cookie. etc. I am up for anyway of doing this cookie idea is just an example So what do you think would be the best / most efficient way of managing this?

    Read the article

  • Javascript undefined behavior with string.replace

    - by epochwolf
    I've been messing around with string.replace and I noticed something very odd with Webkit and Firebug's javascript consoles. I can repeat this behavior in a blank browser window. (Look at the first and last lines) >>> "/literature?page=".replace(/page=/i, "page=2") "/literature?page=" >>> "/literature?page=".replace("page=", "page=2") "/literature?page=2" >>> "/literature?page=".replace(/page=/, "page=2") "/literature?page=2" >>> "/literature?page=".replace(/page=/i, "page=2") "/literature?page=2" Just so nobody thinks I mistyped something, here are screenshots. Firebug (3.0.14) Webkit (Latest nightly as of this post's creation.)

    Read the article

  • ASP.NET MVC ,Maintaining Model State between Ajax requests

    - by Podders
    problem: On first full page request, my controller invokes an applicationServices Layer (Web Service Proxy to my business tier) in order to populate a collection of current services that is stored in my own controller base class property. This is then to be displayed within a view. Everything within the context of that controller has access to this "Services Collection". Now when i make further calls to the same action method via an AJAX Call, i obviously hitt a different instance of that controller meaning my services collection is empty. So other than re-getting the whole collection again, where would i store this collection so it gets persisted between ajax requests? Should i persist it as a seperate DomainModel Object, Session object?....as ViewData is not working for me obv. Excuse my MVC ignorance :) Any help would be greatly appreciated :)

    Read the article

  • How to change type of information for a Title column in SharePoint MOSS 2007 List?

    - by Ruba
    I created a calendar in SharePoint MOSS 2007 that is connected to my Outlook. I added a custom column “Person” to this list and the type of information in this column is: Person or Group. In SharePoint I can hide Title column and in Calendar View show this Person field as Month View Title. So I can see on the calendar who is working that day. Problem is in Outlook. It seems like Outlook doesn’t know a thing about custom fields. In Outlook I can see only Title and few other fields. I could rename Title field to Person, but I can’t change type of information that it contains. By default it is text field and no way to change it to Person or Group. If I could change those “default” column types, then I think my problem would be solved. I know it is possible. I created a custom list, but this list has also those “sticky” Title, Created By and Modified By columns that can’t be changed or removed. Maybe I have to create a custom list with some other program or code? Please help! Thanks in advance!

    Read the article

  • Making footer image stick to bottom of web browser or page

    - by Slyphee
    Hello, I know this has been asked alot of times in the past but for the life of me I can't seem to get any of the other solutions to work. What I'm trying to do is to get the footer (which is an image that repeats across the width of the page) to stick to the bottom of the browser when there isn't enough content to naturally push it to the bottom of the page and when there IS enough content to push it to the bottom it does just that. An example is the one at http://www.themaninblue.com/experiment/footerStickAlt/good_example_short.htm which does exactly what I want but I can't get to work either. The code that I've currently got implemented makes the footer stick to a certain section of the page with text going under it. You can see it at sourcectrl.co.uk but its not much to look at. Heres the code for your viewing pleasure. html, body { font: 100% Arial, Helvetica, sans-serif; height: 100%; color: #597347; margin: 0; padding: 0; background: #573909; } header { display: block; position: relative; top: 0; left: 0; width: 100%; height: 66px; background: url(../images/FillerPage_01.gif) repeat-x left bottom; } section { width: 940px; margin: 0 auto; font-size: 1.4em; overflow: auto; min-height: 100%; height: auto !important; height: 100%; margin-bottom: 87px; position: relative; padding-bottom: 90px; } footer { display:block; position: absolute; bottom: 0; width: 100%; height: 87px; background: url(../images/FillerPage_08.gif) repeat-x left bottom; } Sorry if it seems messy! I'd just like to know if I'm heading in the right direction or theres something I'm just not getting? Oh yeah I'm trying to do all of this with the html 5 markup which is why there is no #footer and the like (could this be why none of the solutions work?). If anyone could give me any help or guidance I'd be soooooo grateful.

    Read the article

  • Using PHP, how to parse the title and meta tags from a HTML page?

    - by Dylan Taylor
    Hey guys, I need to be able to get the TITLE and DESCIPTION metadata out of a page. I've been trying to do this but I've been getting more errors than actual results. (I have an array of about 10 URLS, usually only about 2 of them give me the descrption. I have yet to get the title). So how do I, in PHP, get the Desc and Title from a remote page, and if there is none or if there's an error, ignore it? -Dylan

    Read the article

  • Apache is running but there is no page displayed

    - by Michael Ozeryansky
    I am on a Mac OSX and I am using the built in PHP and Apache2. I have been setting up MySQL and finally when I got MySQL working my local site won't display. Do note that I did have the web server running and delivering PHP enabled pages, just no database connection. But my question is not about MySQL. I have changed various settings in the 'http.conf' file, and I have the line: '127.0.0.1 localhost' in my hosts file. I also have other alias' pointing to 127.0.0.1. I have checked everything I could about Apache and I have made sure that any message in the error_log is ok. I currently have my errorLevel set to debug, so I get all the messages. At this point (HOURS of self fixing) I think I need help. What can I provide for someone to figure this out with me. Thanks.

    Read the article

  • ASP.NET Lifecycle and long process

    - by user204588
    Hello, I know we need a better solution but we need to get this done this way for right now. We have a long import process that's fired when you click start import button on a aspx web page. It takes a long time..sometimes several hours. I changed the timeout and that's fine but I keep getting a connection server reset error after about an hour. I'm thinking it's the asp.net lifecycle and I'd like to know if there are settings in IIS I can change to make this lifecycle last longer.

    Read the article

  • Sending message to windows service by web page

    - by Enriquev
    Hello, How could I do this with no access denied problem? I have a windows service: protected override void OnCustomCommand(int command) { if (command == 1) { foreach (Process traceProcess in Process.GetProcessesByName("notepad.exe")) { traceProcess.Kill(); } } } when I do this: ServiceController sc = new ServiceController("ProjectManager"); if (sc != null) sc.ExecuteCommand(1); From a windows forms it works, but not from a web page, I get access denied on sc.ExecuteCommand. What's the best way for a web page to talk to a service?

    Read the article

  • Quickly check which part of the page lifecycle a control is in

    - by Khanzor
    Is there any way to check what events have fired during the asp.net webforms page/control lifecycle? I know that I can manually add handlers for each event, but that seems a bit ... inefficient. Is there a visualiser, or a property that I can check that will tell me whether these events have fired? EDIT The reason I want to know this is that I am overriding the ViewState property of a custom control, and the viewstate disappears at some point, and I'd like to know at which point in the page lifecycle it is being overriden.

    Read the article

  • Google-Bot fell in love with my 404-page

    - by 32bitfloat
    Every day my access-log looks kind of this: 66.249.78.140 - - [21/Oct/2013:14:37:00 +0200] "GET /robots.txt HTTP/1.1" 200 112 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" 66.249.78.140 - - [21/Oct/2013:14:37:01 +0200] "GET /robots.txt HTTP/1.1" 200 112 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" 66.249.78.140 - - [21/Oct/2013:14:37:01 +0200] "GET /vuqffxiyupdh.html HTTP/1.1" 404 1189 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" or this 66.249.78.140 - - [20/Oct/2013:09:25:29 +0200] "GET /robots.txt HTTP/1.1" 200 112 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" 66.249.75.62 - - [20/Oct/2013:09:25:30 +0200] "GET /robots.txt HTTP/1.1" 200 112 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" 66.249.78.140 - - [20/Oct/2013:09:25:30 +0200] "GET /zjtrtxnsh.html HTTP/1.1" 404 1186 "-" "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" The bot calls the robots.txt twice and after that tries to access a file (zjtrtxnsh.html, vuqffxiyupdh.html, ...) which cannot exist and must return a 404 error. The same procedure every day, just the unexisting html-filename changes. The content of my robots.txt: User-agent: * Disallow: /backend Sitemap: http://mysitesname.de/sitemap.xml The sitemap.xml is readable and valid, so there seems to be no reason why the bot should want to force a 404-error. How should I interpret this behaviour? Does it point to a mistake I've done or should I ignore it?

    Read the article

  • problem with css showing the divisions of the page

    - by user225269
    This is most likely the culprit for this page: http://www.mypicx.com/04192010/ff/ table { border-collapse:collapse; } table,th, td { border: 1px solid black; font-family:Arial, Helvetica, sans-serif; letter-spacing: 1px; font-size:3; font-size:10px; } th { background-color:#A7C942; color:white; } I don't know why, but the page has some php scripts in it which shows a table from mysql database. As you can see in the url above. Please help, why do I see those lines everytime I put those css scripts

    Read the article

  • How to run a WebForms page and an MVC page in different files?

    - by Erx_VB.NExT.Coder
    when i try to do this and load the webforms page, i get this error, even tho the path is correct. what can i do to get past this? i've tried running the aspx page from the root as well. nada. Server Error in '/' Application. The resource cannot be found. Description: HTTP 404. The resource you are looking for (or one of its dependencies) could have been removed, had its name changed, or is temporarily unavailable. Please review the following URL and make sure that it is spelled correctly. Requested URL: /Views/Home/FileUploadFrame.aspx Version Information: Microsoft .NET Framework Version:4.0.30128; ASP.NET Version:4.0.30128.1

    Read the article

  • AJAX ModalPopup Pops Behind (Under) Page Content (Negative z-index)

    - by Aaron Hoffman
    I am having an issue with the AJAX ModalPopupExtender in version 40412 of the AJAX Control Toolkit (http://ajaxcontroltoolkit.codeplex.com/releases/view/43475). The first time the ModalPopup is made visible it works correctly. The z-index is set to 6001 (and the background Div's z-index is set to 6000) and the Popup appears above everything else. If the cancel button within the ModalPopup is clicked, it also has the correct functionality, the display is set to "none" and the ModalPopup is no longer visible. However, when the Popup is triggered again, the z-index is only set to 2000 which is still visible above everything else, but if it is canceled and triggered again it is set to -2000 which is not visible (the z-index is decreasing by 4000 each time). I'm not sure why this is happening. Any ideas how to fix it? Special circumstances: There are multiple ModalPopup's on the page. All ModalPopups are triggered in code-behind through partial-page postbacks (using the .Show() method) ModalPopupExtenders are within the same UpdatePanels that are displayed as popups

    Read the article

  • Measuring Web Page Performance on Client vs. Server

    - by Yaakov Ellis
    I am working with a web page (ASP.net 3.5) that is very complicated and in certain circumstances has major performance issues. It uses Ajax (through the Telerik AjaxManager) for most of its functionality. I would like to be able to measure in some way the amounts of time for the following, for each request: On client submitting request to server Client-to-Server On server initializing request On server processing request Server-to-Client Client rendering, JavaScript processing I have monitored the database traffic and cannot find any obvious culprit. On the other hand, I have a suspicion that some of the Ajax interactions are causing performance issues. However, until I have a way to track the times involved, make a baseline measurement, and measure performance as I tweak, it will be hard to work on the issue. So what is the best way to measure all of these? Is there one tool that can do it? Combination of FireBug and logging inserted into different places in the page life-cycle?

    Read the article

  • jsTree causes that links on whole page are broken

    - by niao
    Greetings, I have the following problem. In my asp.net mvc page (which is a partial view) I create an instance of jsTree as follows: <script type="text/javascript"> $(function() { $("#industries").tree({ callback: { onselect: function(NODE, TREE_OBJ) { $("#SelectedIndustryROWGUID").val($(NODE).attr("id")); $("#resultMessage").append($(NODE).attr("rel")); } }, data: { type: "json", async: true, opts: { method: "GET", url: "/CreateMessage/GetIndustries/" } } }); }); this works fine but then, when I click on any link on the page, it does not work. The links are executed when I choose "open in new tab" option from context menu. Can someone please help me with this?

    Read the article

  • database security with php page that spits out XML

    - by Rees
    Hello, I just created a PHP page that spits outs some data from my database in an XML format. This data is fetched from a flex application I made. I had spent a long time formatting my tables and database information and do not want anyone to be able to simply type www.mysite.com/page_that_spits_out_XML.php and steal my data. However, at the same time I need to be able to access this page from my flex application. Is there a way I can prevent other people from doing this? Thank you!

    Read the article

  • BSD route(8) MAN page bug

    - by Farseeker
    http://www.freebsd.org/cgi/man.cgi?query=route Route is a utility used to manually manipulate the network routing tables. It normally is not needed, as a system routing table management daemon such as routed(8), should tend to this task. ... BUGS The first paragraph may have slightly exaggerated routed(8)'s abilities. Is this really a "bug", or some developer's attempt at humour?

    Read the article

< Previous Page | 114 115 116 117 118 119 120 121 122 123 124 125  | Next Page >