Search Results

Search found 9960 results on 399 pages for 'iwork pages'.

Page 288/399 | < Previous Page | 284 285 286 287 288 289 290 291 292 293 294 295  | Next Page >

  • XML XHR Request resultin in 0 stauts and empty response text.

    - by deepak
    I had another post for the same problem... I think I put the question in a wrong way.. Let me give more details: i have test.html in my c:\ drive and I have a local webserver runnin in qt, and i have some plugin written to that webserver which will get the request and send some response text "hello". in test.html i m making a xml xhr request which will make a GET request like localhost:8080/test which will return the text "hello" by that plugin. Now if I directly open test.html from C:\ it doesnt work, i mean i get response 4 and status 0, and response text nothing.. but the request is passing through webserver and plugin It works fine, when the test.html is put in the webserver pages directory.

    Read the article

  • [LaTex]: Add in the TOC an included PDF

    - by ILoveMyLatexReport
    In my document I include a PDF using \includepdf[pages=-]{./mypdf.pdf} The problem I'm having is how to add a TOC entry for this pdf. It supposed to be an appendix. I tried adding a new section in the appendix but of course the section name can't be printed on the same page than the included pdf, so the resulting TOC line directs to a wrong page. if I use \addcontentsline I loose the numbering and the page is wrong too because the included pdf actually starts at the next page... I'm a bit lost here so I would really appreciate if someone knows how to do this. note: the pdf I try to include was not generated from LaTex. Thanks in advance.

    Read the article

  • Post to a page wall that I'm not admin

    - by Jirico
    I created an app to post to pages the user chooses. How can I direct the message to a page Wall on behalf of the user? From my tests the post is sent, but the page can't see it and the page is not notified either. I have OAuth extended Permissions of publish_stream, what else do I need to my post not being invisible? I will try to clarify the problem in details: 1- I send a post to a Facebook page that is not mine, I'm just a normal user, via application that ask for publish_stream, read_stream permissions. 2- I verify the page logging as admin but no notification is delivered and I can't see the post on my Wall. 3-If I log as the sender user I can see the post on page Wall, it looks like the post is private and just the sender user can see It. If I try to recover the post via Graph API using the same user it returns false. 4- Only the users who has the app installed can see the post on page wall.

    Read the article

  • How can I ensure that JavaScript inserted via AJAX will be executed after the accompanying HTML (als

    - by RenderIn
    I've got portions of pages being replaced with HTML retrieved via AJAX calls. Some of the HTML coming back has JavaScript that needs to be run once in order to initialize the accompanying HTML (setting up event handlers). Since the document has already been loaded, when I replace chunks of HTML using jQuery's .html function, having jQuery(document).ready(function() {...}); doesn't execute since the page loaded long before and this is just a snippet of HTML being replaced. What's the best way to attach event handlers whose code is packaged along with the HTML it's interested in, when that content is loaded via AJAX? Should I just put a procedural block of javascript after the HTML , so that when I insert the new HTML block, jQuery will execute the javascript immediately? Is the HTML definitely in the DOM and ready to be acted upon by JavaScript which is in the same .html call?

    Read the article

  • WebClient.DownloadString() Not Producing Exact HTML

    - by Ryan Fuentes
    So here's the deal. I'm creating a spider bot for a website that scans all the product pages and records the product data. I'm using C# and the WebClient library to download the HTML string. The site I'm crawling must be specially made because the HTML that is received from WebClient.DownloadString() is different than the HTML that I get when I view the source of the HTML when visiting it on a browser. This seems intentional because the only info I can't get is the price. Does anyone know a workaround for this problem or can anyone explain what is happening? Thanks.

    Read the article

  • How do I load the background image from another page?

    - by bbeckford
    Hi all, I'm creating a page that loads content from other pages using jQuery like this: $('#newPage').load('example.html' + ' #pageContent', function() { loadComplete(); }); That all works fine. Now what I want to do is change the background image of the current page to the background image of the page I'm loading from. This is what I'm doing now but I can't for the life of me get it to work: $.get('example.html', function(data) { var pageHTML = $(data); var pageBody = pageHTML.$('body'); alert(pageBody.attr("background")); }); What am I doing wrong?? Thanks, -Ben

    Read the article

  • Wpf application variable not changing

    - by Hasan Zubairi
    In my application I need some variables on almost all the pages so I declared them in Application.xaml like <sys:String x:key="First">First</sys:String> and on one page I change the value in code like Resources["First"] = "This is First"; on the second page when I call the variable like String f = (string)this.TryFindResource("First"); MessageBox.Show(f); the out put is "First" and not "This is First". I also tried like. Application.Current.Properties["First"]

    Read the article

  • Using Selenium to Determining The Visibility of Elements for Print media

    - by Tom Howard
    I would like to determine if particular elements on a page are visible when printed as controlled by CSS @media rules. Is there a way to do this with Selenium? I know there is the isDisplayed method, which takes the CSS into account, but there is nothing I can find to tell Selenium which media type to apply. Is there a way to do this? Or is there another way to test web pages to make sure the elements you want are printed (and those you don't aren't)? Update: For clarity, there are no plans to have a javascript print button. The users will print using the normal print functionality of the browser (Chrome, FF and IE). @media css rules will be used to control what is shown and hidden. I would like Selenium to pretend it is a printer instead of a screen, so I can test if certain elements will be visible in what would be the printed version of the page.

    Read the article

  • How to override JS function from a firefox extension?

    - by BruceBerry
    Hello, I am trying to intercept calls to document.write for all pages. Setting up the interception inside the page by injecting a script like function overrideDocWrite() { alert("Override called"); document.write = function(w) { return function(s) { alert("special dom"); w.call(this, wrapString(s)); }; }(document.write); alert("Override finished"); } is easy and works, but i would like my extension to setup the interception for each document object from inside the extension. I couldn't find a way to do this. I tried to listen for the "load" event and set up the interception there but it also fails. How do I hook calls to doc.write from an extension?

    Read the article

  • c# creating a file from input stream

    - by daemonkid
    My component will recieve a pdf file as a filestream from which I will need to create a file. For testing purposes I am trying to read a file using the filestream object and recreate it at a different location. But the recreated file is created blank. the recreated file has the same number of pages though... This is the code StreamReader sr = new StreamReader(_filePath); str = sr.ReadToEnd(); File.WriteAllText(@"C:\recreated.pdf", str); what am I doing wrong? Thanks for your time.

    Read the article

  • REST client website login

    - by Jordan
    I have written a REST service that uses WSSE as an authentication method but i want to be able to use this rest service through a browser by creating a website around the service. I want the user to be able to log in on the website then when they view, for example the "view users" page an ajax request is made to test.com/users and back comes the list. The part i'm trying to get my head around is the logging in/out on the website and keeping the user logged in across pages. Since in a true REST implementation there's no state held on the server, i can't use $_SESSION and now i don't know where to start! What is the best way to go about this? Do i still need to store session information on the server then possibly use cURL to make the request? Thanks Jay

    Read the article

  • how to get apache mod_cache work with mod_wsgi (django)?

    - by harmv
    I thought i'd speed up my django projects, by letting apache doing some caching for me. Unfortunately I see that apache never caches my dynamic pages. Has mod_cache problems with mod_wsgi served code ? My apache config: <VirtualHost *:80 ServerName myserver.com CacheEnable mem / # for testing only CacheIgnoreQueryString On CacheIgnoreCacheControl On WSGIDaemonProcess aname processes=1 threads=25 WSGIProcessGroup aname Alias /media/ /home/harm/projects/test/media/ WSGIScriptAlias / /home/harm/projects/test/wsgi.py The response does have the correct caching headers: Content-Length 2647 Content-Encoding gzip Vary Accept-Encoding Cache-Control public, max-age=3600 Keep-Alive timeout=15, max=100 Connection Keep-Alive Content-Type application/x-javascript Am I missing something ?

    Read the article

  • Are SqlCipher open cursors a security concern?

    - by user1178479
    I'm using SqlCipher with content providers. Right now, when I want to lock the app I just clear out the cached password. However, the app can continue to work with any open cursors. This means that re-opening the app grants access to the sensitive data. I fix this issue on the surface by redirecting to a login screen if the app doesn't have passwords. However, I'm concerned if there are any security issues with these open cursors or if I should just continue to block UI access and not worry? SqlCipher's docs say that it reads/writes encrypted pages on the fly, as opposed to decrypting the entire DB, this makes me think that open cursors are still secure. The main concern here is that someone loses their phone and then a knowledgeable individual can use these open cursors to extract sensitive data.

    Read the article

  • mySQL php query - news/ friends feed

    - by rpsep2
    I want to show a user the recent uploads from their friends. I have the users friends id's in an array: $friends A user could have, potentially, thousands of friends. I can select the uploads from 1 of a users friends with: $row = $mysqli->query("SELECT * FROM photos WHERE uploader_id = ".$friend." ORDER BY date_uploaded DESC LIMIT ".$page.", 25"); But I need to find all of a users friends uploads. I thought about doing this in a loop iterating over the $friends array, but then I'd be potentially running thousands of mysql queries. How can I do this most efficiently? so to clarify: search a 'photos' table for photos which are uploaded by specific users(friends), held in $friends variable, sort by date_uploaded and limit to x results so I can have pages 1, 2, 3 etc.

    Read the article

  • Don't Change URL in Browser When Clicking <asp:LinkButton>

    - by Corey Goldberg
    I have an ASP.NET page that uses a menu based on asp:LinkButton control in a Master page. When a user selects a menu item, an onclick handler calls a method in my C# code. The method it calls just does a Server.Transfer() to a new page. From what I have read, this is not supposed to change the URL displayed in the browser. The problem is it that the URL changes in the browser as the user navigates the menu to different pages. Here is an item in the menu: <asp:LinkButton id="foo" runat="server" onclick="changeToHelp"><span>Help</span> </asp:LinkButton> In my C# code, I handle the event with a method like: protected void changeToHelp(object sender, EventArgs e) { Server.Transfer("Help.aspx"); } Any ideas how I can navigate through the menu without the browser's URL bar changing?

    Read the article

  • Wordpress auto-generated "canonical" link - get them to use a custom URL parameter?

    - by kiko
    Hello - Does anyone know how to modify the Wordpress canonical links to add a custom URL parameter? I have a Wordpress site with a page that queries a separate (non-Wordpress) database. I passed the URL parameter "pubID" to display individual books and it is working OK. Example: http://www.uglyducklingpresse.org/catalog/browse/item/?pubID=63 But the individual books are not showing up properly in Google - the ?pubID parameter is stripped out. I think maybe this is because all the item pages have the same auto-generated "canonical" URL link tag in the source - one with the "pubID" parameter stripped out. Example: link rel='canonical' href='http://www.uglyducklingpresse.org/catalog/browse/item/' Is there a way to perhaps edit .htaccess to add a custom URL parameter to Wordpress, so that the parameter is not stripped out by permalinks and the "canonical" links? Or maybe there's another solution ... Thank you for any ideas!

    Read the article

  • Anonymous users support vs Google bot

    - by Andy
    I have a User class in my web app that represents a user currently logged in. Every time a user vists a page, a User instance is populated based on authentication data supplied in cookies. A User instance is created even if an anonymous user logs in - and a corresponding new record is created in the User table in the database. This approach allows me to save some state info for the current user regardless of its type. The problem however with this approach is the Google bot, and other non-human web organisms crawling my pages. Every time a bot starts to walk around the site, thousands of useless records will be created in the database, each of them only to be used for a single page. Question: what is the best trade off? How to support anonymous users, save their state, and don't get too much overhead because of cookieless bots?

    Read the article

  • RSS Reader php (have already read related articles)

    - by lightingwrist
    Hey there. I've read all the related articles on here and can't find one that is specific to what I am looking for. I am new to RSS and am looking for the following reader if anyone know's the right direction to throw me in: An rss reader that I can put on my page that does NOT require mysql database A fairly light chunk of code that I can just add as many .xml,rss.php links/addresses to I can wrap div's around to style each segment specifically as possible can manually limit the amount of feeds that are read to conform to my desires of the pages content out put thanks in advance!

    Read the article

  • BufferedReader ready method in a while loop to determine EOF?

    - by BobTurbo
    I have a large file (wikipedia english arcticles only database as xml file) I am using to read one character at a time using BufferedReader. The psuedo code is: file = BufferedReader... while (file.ready()) character = file.read() is this actually valid? Or will ready just return false when it is waiting for the HDD to return data and not actually when the EOF has been reached? I tried to use if (file.read() == -1) but seemed to run into an infinite loop that I literally could not find. I am just wondering if it is reading the whole file as my statistics say 444 380 wikipedia pages have been read but I thought there were many more articles..

    Read the article

  • Remove objects from different environments

    - by Fred
    I have an R script file that executes a second R script via: source("../scripts/second_file.R") That second file has the following lines: myfiles <- list.files(".",pattern = "*.csv") ... rm(myfiles) When I run the master R file I get: > source("../scripts/second_file.R") Error in file.remove(myfiles) : object 'myfiles' not found and the program aborts. I think this has something to do with the environment. I looked at ?rm() pages but less than illuminating. I figure I have to give it a position argument, but not sure which.

    Read the article

  • Paging a UIScrollView with a large PDF

    - by Fousa
    I try to create a simple UIScrollView with paging. And I want to be able to scroll through a large PDF document, but this gives me some problems... I tried the following options: Convert all the PDF pages to UIImages at startup, this works, but is very slow on start Manually drawing the PDF page in the drawRect, but yet again this was slow... And I prefer not to load everything at startup but to do it during the usage. Did anyone did this recently? Can't seem to find a nice example project. Thnx! Jelle

    Read the article

  • trigger click behaviour of image-map's area

    - by Amit
    I have a image map set up and each area in the image map has a href defined. the href on area contains urls to other pages in my application. i generate a have a small ul which lists down name attribute of the area tag. i want the dynamically generated ul/lis to imitate click behaviour of area tag. for this, i have the following jquery set up - $('li').click(function(e){ $('area[name='+$(this).html()+']').trigger('click'); }); but the above works well only in ie6+. ff does not fire the click event. i also tried the click() variant but to no avail. looking forward for some help. Thanks :)

    Read the article

  • call function from external js (php) file

    - by bah
    Hi, I have a file where I keep all scripts so my pages wouldn't get messy (I have php file which generates required javascript). My includes basically look like this: </html> <head> <script type="text/javascript" src="jquery.js"></script> <script type="text/javascript" src="script.php"></script> </head> <body> <input type="button" onClick="return blah();" /> </body> </html> In script.php there's jquery wrapper $(document).ready where I keep all jquery related stuff. The funny thing is, when I put function blah() inside this wrapper I get "blah is not defined" error, but when I put it outside - works perfectly. So, what could be the problem?

    Read the article

  • Best method of Zend Framework caching

    - by iamthejeff
    I have a blog built using Zend Framework, which I realize might be a bit overkill for a blog alone, but I am planning on adding other features in the future. Nevertheless, I've noticed pages could be a little speedier. I've done a basic caching method that basically captures everything in index.php (Core frontend and File backend), which works great, but unfortunately it also prevents dynamic page contents from updating (messages like "this was posted 5 minutes ago", etc) until the cache period expires. So my question is what would be the best method of caching to improve performance? I am doing fairly basic queries which are mostly simple selects, not many joins or anything fancy (using Zend_Db_Table), and even on a small database page loads are a little sluggish. Is it worth it to cache queries or should I focus my time elsewhere?

    Read the article

  • best way to prgram php and mysql query

    - by air
    most of times in our pages we need to print only 1 field value of table in loop. for example <?php for($i=1;$i<=mysql_num_rows($result);$i++) { echo $row['name']; $sql1="select industry from table_industry where profid='".$row['prof']."'"; $result1=mysql_query($sql1); $row1=mysql_fetch_array($result1); echo $row1['industry']; ?> } ?> above is one PHP code just for example where we think that $row['prof'] carry value of profession ID and we print profession of each person. for example we have 5000+ record in table and above loop will execute for 5000+ times what will be the best way to print value of industry field from table table_industry Same code which i write above ? any other php code suggestion for faster execution and less use of resources? Thanks

    Read the article

< Previous Page | 284 285 286 287 288 289 290 291 292 293 294 295  | Next Page >