Search Results

Search found 59112 results on 2365 pages for 'web internet'.

Page 13/2365 | < Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >

  • Cannot connect to Internet after using Windows XP on dual booted machine

    - by babelproofreader
    I have a dual booted Windows XP and Ubuntu 10.10 box (AMD64) and have an unusual problem in that after connecting to the internet on Windows I cannot connect using Ubuntu. I don't believe it is actually a problem with configuring my Ubuntu connections because when I get up in the morning Ubuntu connects without any problem - this problem only exists (and persists for at least several hours) if I try to connect using Ubuntu immediately after connecting using Windows. I don't know how long the problem actually lasts because I usually just go to bed and then in the morning I can connect without problems. It is not problem with my provider's service as I have a second Ubuntu 10.10 box which I manually connect and works even when the dual booted one (on Ubuntu) won't. It is a particular problem with the dual booted box. A power down reset does not solve the problem.

    Read the article

  • Wireless internet sharing stops working after client disconnects

    - by Mandy
    I'm running Unbuntu 11.10 with the latest service packs, sharing an ether connection over wireless. Everything works fine, up to the point that the client (a laptop) sleeps or otherwise disconnects. The wireless connection at that point no longer functions, not showing up when searching for the hotspot with the laptop. The only way to recover is to reboot the machine, or toggle the wireless share off/on. I've double checked the wireless card has power management off. Currently what I'm doing is keeping an Ipod plugged into the box 24/7 to keep the wireless connection up when the laptop disconnects. I've searched the internet far and wide, found many issues with wireless problems after an ubuntu box itself sleeps, but nothing directly related to my problem. Hoping someone has some idea of where to look and what to check. A girl needs her Ipod back!

    Read the article

  • Mts mblaze doesn't connect to internet

    - by Chetan
    I've sony vaio vpceb24en laptop, I'm using dual boot system with windows 7 and ubuntu. Firstly I installed ubuntu 10.04 LTS then upgraded to 12.04 & 12.04.1 afterwards. I have mts mblaze dongle. I have set up the mts connection profile under the mobile broadband. My problem is : Sometimes my connection is easily made & I use it, sometime it never connects. So I have to restart my system then it connects to internet sometimes. But now it's not connecting even I restarted my system many times, just it shows trying to connect icon once, then never tries to connect. while in windows it's working fine. Please someone help me out how to resolve this problem as I'm very new to linux. Thanks, Chetan

    Read the article

  • 12.10 unable to connect to the Internet using Chrome or Firefox

    - by Thomas Anderson
    I received a software update for Ubuntu 12.10 while using Zorin OS6 wireless. I accepted the update and installed 12.10. Please tell me the step by step, using layman's terms, on how to get Google Chrome, or Mozilla Firefox to connect to the Internet? I am receiving DNS server error with both browsers. I can connect to my router wired or wireless. I posted and received initial information but was not sure exactly what some of the terms were and or what to do next etc. I am not necessarily familiar with some technical terms, or do I need to install additional plug-ins? I do not know what proxy or DCHP means or what do with these terms please, please help. I am not a programmer, nor do I work in IT. Again please help.

    Read the article

  • Connman not connecting to internet

    - by sagarchalise
    I wanted to try connman as I heard it will be replacing network manager. I kinda liked its looks and feel but the problem is I donot get connected to internet. I am using ADSL and it connects to my adsl router but doesnot ping google or any other sites. I tried to change the default gateway as well as use static ips but to no avail. It displays I am wired but doesnot open any sites or ping to any external sites except for 198.168.1.1 which is my router ip. Can anyone help ? Basically it is doing alright with network-manager as well as wicd

    Read the article

  • Internet connection randomly drops out

    - by Dr_Bunsen
    Since a week or 2, I have this random network failures. No error pops up, the system says it's still connected to the internet(wifi and wired). The problem is, I can't connect to anything, even a ping google.com gives no feedback. And after many network-manager resets it will work again, for just a minute, and drop out again. I noticed that the /etc/resolv.conf contains this line: nameserver 127.0.0.1 I left out the "Do not edit" stuff, and after it worked for a second, this line is added: search lan When I change the nameserver to 8.8.8.8, I am still able to connect to anywhere. Does anyone got a clue what is happening here? I will provide any information asked. Thanks in advance p.s. I run Ubuntu 13.04 since lastnight(24-06-2013)

    Read the article

  • No internet access since 12.04

    - by Roelof
    I have updated yesterday to Ubuntu 12.04 but can not access the Internet. It is an old PC without wireless. But the older Ubuntu I had installed up until now (11.??) worked perfectly. I'm not that familiar with Ubuntu, and I have been trying Google for some time know, but I cannot find the answer. I noticed that running nm-tool resulted in some things that don't sound right: Type: wired Driver: ne2k-pci state: unavailable Default: no Carrier: off Does anyone have a way to fix the connection?

    Read the article

  • How to connect to the internet through broadband through ALFA USB adapter

    - by user91045
    Hi evryone I have question that really make me mad..I search every where and I couldn't find the answer.pleas read the details carefully: I have broadband (Wan mini-port PPPOE) account from a wierless network near from us...I'm receiving the signal with grid and Alfa AWUS036H wireless usb adapter.In all of windows version I can connect through my broadband when I do these steps 1.setup new connection or network 2.connect to the internet 3.setup new broadband connection then I but the ISP name(not Important just to remember my broadband connection) and then i put the username and password then connect. I can find and connect to my wireless networkbut when I make Dsl connection in ubuntu 12.04LTS and I put the username and password I couldn't connect.. I tried this tool"gnome dial up tool" and wouldn't work.. in other way..(how can I connect to broadband connection through ALFA USB adapter??) my wireless network have a servers when i check with winbox: mac adress ip adress server name version router board verson and I connect PPP through them... any answer will help me ..so we could solve this problem. thanks.

    Read the article

  • No GUI, No internet connection, please help 12.04

    - by KB_
    I am new to Ubuntu and I tried to install ubuntu server 12.04 on my laptop. now my problem is - I am unable to connect to the internet (I have wifi ONLY connection). ubuntu didn't recognize my built-in wifi on my Toshiba Satellite L505. There is no GUI. I have Terminal only, I tried Sudo apt-get update but i am getting errer msg because of no connection. 1. I need to know if there is any possible way that i can download and install driver for my wifi. 2. what other option do i have to be able to update ubuntu. Thanks KB

    Read the article

  • System crashes when internet connection is unplugged

    - by Rincewind
    When i first tried to install any ubuntu newer than 10.04 on my netbook (Acer aspire one 722 11,6" WXGA - C-60 Dual Core APU - AMD Radeon HD 6250 - 1x HDMI - 2GB DDR3 - 320GB - 802.11bgN - 6CELL2.2 - BT 3.0 - VGA Webcam), it would fail to boot. I found out that the only fix to this was linking it to the internet via lan-cable during the installation. However, after installing ubuntu 12.04 successfully my system is crashing everytime i unplug the lan-cable. My first thought was disabling the netbook´s w-lan (via the band at the top) but it was not fixing the problem. Looking on the taskmanager ('top') during the process brought no further insight, no process was exploding or something like that. As i am out of guesses help would be really appreaciated. Lg Rincewind

    Read the article

  • Search the internet for a term selected in a text editor/viewer

    - by cipricus
    It seems possible (as I learned from a comment to this question made by user55822) to use Artha in order to search a term on the Internet by using "keybinding". What the aforementioned user says is that "you could search mouse selection anywhere by using an app like Artha, or by keybinding sh -c 'firefox "http://translate.google.com/#en/fr/$(xclip -o)"' (example given translate english into french)" A lot of things are obscure for me here. I could use Artha or keybinding (as alternative), or both? How to do that in each case? and "The command I gave you will open google translate in Firefox with your mouse selection (text highlighted) as a query. Use sh -c 'firefox "https://www.google.com/search?q=$(xclip -o)"' if you want regular Google over Google translate. And obviously, that would work everywhere" I do not know what to do with that formula. What should I do exactly? How to use it? In what application?

    Read the article

  • Cannot connect to Internet on 11.04 using BSNL EVDO Prithvi Card

    - by Joy
    I cannot connect to Internet using BSNL EVDO Prithvi data card. Went through some websites that offered help, installed wvdial package and tried again, but was unsuccessful. I have Read that, Ubuntu 11.04 automatically detects Data Card, You only need to configure "Network Manager" and it will work, I did exactly that, but the result is same. The OS detects the data card, and the presence of network , but it cannot login. I have read in some forums that Ubuntu 11.04 does not have support for BSNL EVDO Prithvi, is it true? I re-checked the "User ID" and "Password". Its working on Windows. Please help me fix this.

    Read the article

  • how to add programs to ubuntu without internet access

    - by captainandcoke
    I don't have internet access at my home and it takes me about a half hour to ride my bike to the library. I have downloaded .deb files to try to install at my home computer but everyone I have downloaded says it can't install because it depends on package X. The next day I will download package X and it will require package Y. Is there anyway to find out what ALL the sub-dependencies are for deb files? I have tried to boot from USB or External Hard drive on the library computers but the security settings prevent this. Also, I do not know anyone with a Linux computer.

    Read the article

  • Unable to connect to internet

    - by Peter Schriver
    I just upgraded to 12.10. Everything appears to go well but I am unable to connect to the Internet. The LAN is fine and I can view my shared files but no e-mail or access to web pages. My other computer which is running 12.04 or older seems to connect fine. I only upgraded one to make sure things would be ok. I am currently connecting fine with a wireless Windows laptop. Output from cat /etc/resolv.conf # Generated by NetworkManager nameserver 127.0.0.1

    Read the article

  • Internet Connection Problems

    - by Confused One
    I just installed Ubuntu 11.04 onto a laptop. There is no list of automatic connections when I click on the internet connection symbol. Enable Wireless was unchecked, so I checked it. However, it continues saying "wireless is disabled." Soft is blocked, and when I tried rfkill unblock wifi it told me that the device was not ready. I tried to counter this with what I found on this thread but nothing worked. I have tried manually entering my Wireless information, but this does not help. Is there a solution?

    Read the article

  • Problem hosting WebBrowser control in an ATL app.

    - by Andrew Bucknell
    I have a legacy atl app that hosts a webbrowser control in an ATL window. I create an instance of the client to host the browser using the following sequence CComPtr<IOleObject> spOleObject; HRESULT hr = CoCreateInstance(CLSID_WebBrowser, NULL, CLSCTX_INPROC, ID_IOleObject,(void**)&spOleObject); spOleObject->SetClientSite(this); GetClientRect(&rcClient); hr = spOleObject->DoVerb(OLEIVERB_INPLACEACTIVATE, &msg, this, 0, m_hWnd, &rcClient); hr = AtlAdvise(m_spWebBrowser, GetUnknown(), DIID_DWebBrowserEvents2, &m_dwCookie); CComVariant navvar(navurl); m_spWebBrowser->Navigate2(&navvar, NULL, NULL, NULL, NULL); This sequence works fine to create the initial browse window. The call to navigate2 works and if I look at the window via spy++ I have Shell Embedding - Shell DocObject View - Internet Explorer_Server. When a popup occurs (detected through NewWindow3) I launch a new window and execute the same code sequence for the new window. In the popup window the navigate2 doesnt work, and when I look at this new window in spy++ I just have Shell Embedding. I get the same problem even if I instantiate the popup window on startup, so its not related to NewWindow3 at all - it seems the second instance of the web control isnt instantiating even though all the calls return S_OK. This sequence worked fine under IE7 but now I am using IE8 and the popup window isnt working. There is clearly something I am missing but I cant guess what it may be. Any suggestions would be incredibly helpful.

    Read the article

  • Internet Explorer 7 Bugs - incorrect display OR dead links

    - by ClarkeyBoy
    Hi, I recently launched a website I have been developing over the past year - http://Live.heritageartpapers.co.uk/. My dad, who owns the company, had a phone call today saying it doesnt display properly in IE7. Bug #1: The header and footer are both in a div, whereas the content is in a table between the two divs. Reportedly the content (table) sometimes (not always, according to IETester) displays below the footer, but the footer still displays where it is supposed to (ie there is a massive gap where the content should fit). Bug #2: When the content does display in the correct place, all the links on the page are dead - click on them and nothing happens. As you can see if you view it in Firefox (the version I am using is 3.6), the links in the left hand menu turn orange on mouseover. However they do not even do this in IE7. Note that they do turn orange and do work if the content is displayed below the footer. I cant see why its happening - according to IETester, the IE7 interpreted source code has all the tags capitalised and many quotes removed (for example for the id attribute on most, if not all, tags) but I doubt this could cause the above bugs, could it? My question is whether anyone has ever seen any of these problems before, and/or has a solution to any of these problems?? I currently do not have the application open, but will post any relevant code in a few minutes. Alternatively just use view source. Many thanks in advance. Regards, Richard Clarke

    Read the article

  • Why are these floats wrapping in Internet Explorer?

    - by Cord Blomquist
    My website, ReadyMadeWeb.com, looks great in FireFox, Safari, and Chrome, but of course IE has different plans for the floats used the in the #content area. You can see what I mean if you view the site in IE. The main column is pushed below the first sidebar. The content boxes and the de.licio.us sidebar both sit within another div. I have looked at overflow and clear settings, but changing those has not helped. Any ideas?

    Read the article

  • Single click to activate last tab in windows 7 internet explorer (like FF)

    - by Kirk Hings
    Windows 7 annoyance: the jump lists are fine for hovering over the taskbar items, great go ahead and show me previews of all the open instances and maybe I'll pick the exact one I want. However this works different between IE and FF. Hovering over the FF icon just shows me the single FF window, not any of the tabs I have open. Clicking it means my FF window opens and I'm at the last tab I was on. Great, normal behavior, one click and I'm in. Hovering over the IE icon shows me all the tabs in my single IE window. Ok, but I just want to go back to my last active IE tab. I can't! I am forced to click twice to go back to any IE tab. Sure it's a minor annoyance, but as I'm developing throughout the day I find it more and more annoying. Any ideas to fix so I can just click the IE taskbar icon once and it pulls up the last tab I was on?

    Read the article

  • Focus CSS tag in Internet Explorer 8

    - by Sam
    This is driving me nuts. http://www.cssdrive.com/index.php/examples/exampleitem/focus_pseudo_class This is an example of using the hover pseudo-class. Works fine in Chrome and IE. When I save locally it works fine in Chrome but won't work in IE. What am I doing wrong!? <link rel="Stylesheet" href="style.css" /> <form> <p>1) Name:<br /> <input type="text" size="40"></p> <p>2) Email address:<br /> <input type="text" size="40"></p> <p>3) Comments:<br /> <textarea rows="5" name="comments" cols="45" wrap="virtual"></textarea></p> <p><input id="actualsubmit" type="submit" value="Submit"></p> </form> style.css: input:focus, textarea:focus{ background-color: lightyellow; }

    Read the article

  • Web Safe Area (optimal resolution) for web app design?

    - by M.A.X
    I'm in the process of designing a new web app and I'm wondering for what 'Web Safe Area' should I optimize the app layout and design. By Web Safe Area I mean the actual area available to display the website in the browser (which is influenced by monitor resolution as well as the space taken up by the browser and OS) I did some investigation and thinking on my own but wanted to share this to see what the general opinion is. Here is what I found: Optimal Display Resolution: w3schools web stats seems to be the most referenced source (however they state that these are results from their site and is biased towards tech savvy users) http://www.w3counter.com/globalstats.php (aggregate data from something like 15,000 different sites that use their tracking services) StatCounter Global Stats Display Resolution (Stats are based on aggregate data collected by StatCounter on a sample exceeding 15 billion pageviews per month collected from across the StatCounter network of more than 3 million websites) NetMarketShare Screen Resolutions (marketshare.hitslink.com) (a web analytics consulting firm, they get data from browsers of site visitors to their on-demand network of live stats customers. The data is compiled from approximately 160 million visitors per month) Display Resolution Summary: There is a bit of variation between the above sources but in general as of Jan 2011 looks like 1024x768 is about 20%, while ~85% have a higher resolution of at least 1280x768 (1280x800 is the most common of these with 15-20% of total web, depending on the source; 1280x1024 and 1366x768 follow behind with 9-14% of the share). My guess would be that the higher resolution values will be even more common if we filter on North America, and even higher if we filter on N.American corporate users (unfortunately I couldn't find any free geographically filtered statistics). Another point to note is that the 1024x768 desktop user population is likely lower than the aforementioned 20%, seeing as the iPad (1024x768 native display) is likely propping up those number (the app I'm designing is flash based, Apple mobile devices don't support flash so iPad support isn't a concern). My recommendation would be to optimize around the 1280x768 constraint (*note: 1280x768 is actually a relatively rare resolution, but I think it's a valid constraint range considering that 1366x768 is relatively common and 1280 is the most common horizontal resolution). Browser + OS Constraints: To further add to the constraints we have to subtract the space taken up by the browser (assuming IE, which is the most space consuming) and the OS (assuming WinXP-Win7): Win7 has the biggest taskbar footprint at a height of 40px (XP's and Vista's is 30px) The default IE8 view uses up 25px at the bottom of the screen with the status bar and a further 120px at the top of the screen with the windows title bar and the browser UI (assuming the default 'favorites' toolbar is present, it would instead be 91px without the favorites toolbar). Assuming no scrollbar, we also loose a total of 4px horizontally for the window outline. This means that we are left with 583px of vertical space and 1276px of horizontal. In other words, a Web Safe Area of 1276 x 583 Is this a correct line of thinking? I'm really surprised that I couldn't find this type of investigation anywhere on the web. Lots of websites talk about designing for 1024x768, but that's only half the equation! There is no mention of browser/OS influences on the actual area you have to display the site/app. Any help on this would be greatly appreciated! Thanks. EDIT Another caveat to my line of thinking above is that different browsers actually take up different amounts of pixels based on the OS they're running on. For example, under WinXP IE8 takes up 142px on top of the screen (instead the aforementioned 120px for Win7) because the file menu shows up by default on XP while in Win7 the file menu is hidden by default. So it looks like on WinXP + IE8 the Web Safe Area would be a mere 572px (768px-142-30-24=572)

    Read the article

  • Passing multiple POST parameters to Web API Controller Methods

    - by Rick Strahl
    ASP.NET Web API introduces a new API for creating REST APIs and making AJAX callbacks to the server. This new API provides a host of new great functionality that unifies many of the features of many of the various AJAX/REST APIs that Microsoft created before it - ASP.NET AJAX, WCF REST specifically - and combines them into a whole more consistent API. Web API addresses many of the concerns that developers had with these older APIs, namely that it was very difficult to build consistent REST style resource APIs easily. While Web API provides many new features and makes many scenarios much easier, a lot of the focus has been on making it easier to build REST compliant APIs that are focused on resource based solutions and HTTP verbs. But  RPC style calls that are common with AJAX callbacks in Web applications, have gotten a lot less focus and there are a few scenarios that are not that obvious, especially if you're expecting Web API to provide functionality similar to ASP.NET AJAX style AJAX callbacks. RPC vs. 'Proper' REST RPC style HTTP calls mimic calling a method with parameters and returning a result. Rather than mapping explicit server side resources or 'nouns' RPC calls tend simply map a server side operation, passing in parameters and receiving a typed result where parameters and result values are marshaled over HTTP. Typically RPC calls - like SOAP calls - tend to always be POST operations rather than following HTTP conventions and using the GET/POST/PUT/DELETE etc. verbs to implicitly determine what operation needs to be fired. RPC might not be considered 'cool' anymore, but for typical private AJAX backend operations of a Web site I'd wager that a large percentage of use cases of Web API will fall towards RPC style calls rather than 'proper' REST style APIs. Web applications that have needs for things like live validation against data, filling data based on user inputs, handling small UI updates often don't lend themselves very well to limited HTTP verb usage. It might not be what the cool kids do, but I don't see RPC calls getting replaced by proper REST APIs any time soon.  Proper REST has its place - for 'real' API scenarios that manage and publish/share resources, but for more transactional operations RPC seems a better choice and much easier to implement than trying to shoehorn a boatload of endpoint methods into a few HTTP verbs. In any case Web API does a good job of providing both RPC abstraction as well as the HTTP Verb/REST abstraction. RPC works well out of the box, but there are some differences especially if you're coming from ASP.NET AJAX service or WCF Rest when it comes to multiple parameters. Action Routing for RPC Style Calls If you've looked at Web API demos you've probably seen a bunch of examples of how to create HTTP Verb based routing endpoints. Verb based routing essentially maps a controller and then uses HTTP verbs to map the methods that are called in response to HTTP requests. This works great for resource APIs but doesn't work so well when you have many operational methods in a single controller. HTTP Verb routing is limited to the few HTTP verbs available (plus separate method signatures) and - worse than that - you can't easily extend the controller with custom routes or action routing beyond that. Thankfully Web API also supports Action based routing which allows you create RPC style endpoints fairly easily:RouteTable.Routes.MapHttpRoute( name: "AlbumRpcApiAction", routeTemplate: "albums/{action}/{title}", defaults: new { title = RouteParameter.Optional, controller = "AlbumApi", action = "GetAblums" } ); This uses traditional MVC style {action} method routing which is different from the HTTP verb based routing you might have read a bunch about in conjunction with Web API. Action based routing like above lets you specify an end point method in a Web API controller either via the {action} parameter in the route string or via a default value for custom routes. Using routing you can pass multiple parameters either on the route itself or pass parameters on the query string, via ModelBinding or content value binding. For most common scenarios this actually works very well. As long as you are passing either a single complex type via a POST operation, or multiple simple types via query string or POST buffer, there's no issue. But if you need to pass multiple parameters as was easily done with WCF REST or ASP.NET AJAX things are not so obvious. Web API has no issue allowing for single parameter like this:[HttpPost] public string PostAlbum(Album album) { return String.Format("{0} {1:d}", album.AlbumName, album.Entered); } There are actually two ways to call this endpoint: albums/PostAlbum Using the Model Binder with plain POST values In this mechanism you're sending plain urlencoded POST values to the server which the ModelBinder then maps the parameter. Each property value is matched to each matching POST value. This works similar to the way that MVC's  ModelBinder works. Here's how you can POST using the ModelBinder and jQuery:$.ajax( { url: "albums/PostAlbum", type: "POST", data: { AlbumName: "Dirty Deeds", Entered: "5/1/2012" }, success: function (result) { alert(result); }, error: function (xhr, status, p3, p4) { var err = "Error " + " " + status + " " + p3; if (xhr.responseText && xhr.responseText[0] == "{") err = JSON.parse(xhr.responseText).message; alert(err); } }); Here's what the POST data looks like for this request: The model binder and it's straight form based POST mechanism is great for posting data directly from HTML pages to model objects. It avoids having to do manual conversions for many operations and is a great boon for AJAX callback requests. Using Web API JSON Formatter The other option is to post data using a JSON string. The process for this is similar except that you create a JavaScript object and serialize it to JSON first.album = { AlbumName: "PowerAge", Entered: new Date(1977,0,1) } $.ajax( { url: "albums/PostAlbum", type: "POST", contentType: "application/json", data: JSON.stringify(album), success: function (result) { alert(result); } }); Here the data is sent using a JSON object rather than form data and the data is JSON encoded over the wire. The trace reveals that the data is sent using plain JSON (Source above), which is a little more efficient since there's no UrlEncoding that occurs. BTW, notice that WebAPI automatically deals with the date. I provided the date as a plain string, rather than a JavaScript date value and the Formatter and ModelBinder both automatically map the date propertly to the Entered DateTime property of the Album object. Passing multiple Parameters to a Web API Controller Single parameters work fine in either of these RPC scenarios and that's to be expected. ModelBinding always works against a single object because it maps a model. But what happens when you want to pass multiple parameters? Consider an API Controller method that has a signature like the following:[HttpPost] public string PostAlbum(Album album, string userToken) Here I'm asking to pass two objects to an RPC method. Is that possible? This used to be fairly straight forward either with WCF REST and ASP.NET AJAX ASMX services, but as far as I can tell this is not directly possible using a POST operation with WebAPI. There a few workarounds that you can use to make this work: Use both POST *and* QueryString Parameters in Conjunction If you have both complex and simple parameters, you can pass simple parameters on the query string. The above would actually work with: /album/PostAlbum?userToken=sekkritt but that's not always possible. In this example it might not be a good idea to pass a user token on the query string though. It also won't work if you need to pass multiple complex objects, since query string values do not support complex type mapping. They only work with simple types. Use a single Object that wraps the two Parameters If you go by service based architecture guidelines every service method should always pass and return a single value only. The input should wrap potentially multiple input parameters and the output should convey status as well as provide the result value. You typically have a xxxRequest and a xxxResponse class that wraps the inputs and outputs. Here's what this method might look like:public PostAlbumResponse PostAlbum(PostAlbumRequest request) { var album = request.Album; var userToken = request.UserToken; return new PostAlbumResponse() { IsSuccess = true, Result = String.Format("{0} {1:d} {2}", album.AlbumName, album.Entered,userToken) }; } with these support types:public class PostAlbumRequest { public Album Album { get; set; } public User User { get; set; } public string UserToken { get; set; } } public class PostAlbumResponse { public string Result { get; set; } public bool IsSuccess { get; set; } public string ErrorMessage { get; set; } }   To call this method you now have to assemble these objects on the client and send it up as JSON:var album = { AlbumName: "PowerAge", Entered: "1/1/1977" } var user = { Name: "Rick" } var userToken = "sekkritt"; $.ajax( { url: "samples/PostAlbum", type: "POST", contentType: "application/json", data: JSON.stringify({ Album: album, User: user, UserToken: userToken }), success: function (result) { alert(result.Result); } }); I assemble the individual types first and then combine them in the data: property of the $.ajax() call into the actual object passed to the server, that mimics the structure of PostAlbumRequest server class that has Album, User and UserToken properties. This works well enough but it gets tedious if you have to create Request and Response types for each method signature. If you have common parameters that are always passed (like you always pass an album or usertoken) you might be able to abstract this to use a single object that gets reused for all methods, but this gets confusing too: Overload a single 'parameter' too much and it becomes a nightmare to decipher what your method actual can use. Use JObject to parse multiple Property Values out of an Object If you recall, ASP.NET AJAX and WCF REST used a 'wrapper' object to make default AJAX calls. Rather than directly calling a service you always passed an object which contained properties for each parameter: { parm1: Value, parm2: Value2 } WCF REST/ASP.NET AJAX would then parse this top level property values and map them to the parameters of the endpoint method. This automatic type wrapping functionality is no longer available directly in Web API, but since Web API now uses JSON.NET for it's JSON serializer you can actually simulate that behavior with a little extra code. You can use the JObject class to receive a dynamic JSON result and then using the dynamic cast of JObject to walk through the child objects and even parse them into strongly typed objects. Here's how to do this on the API Controller end:[HttpPost] public string PostAlbum(JObject jsonData) { dynamic json = jsonData; JObject jalbum = json.Album; JObject juser = json.User; string token = json.UserToken; var album = jalbum.ToObject<Album>(); var user = juser.ToObject<User>(); return String.Format("{0} {1} {2}", album.AlbumName, user.Name, token); } This is clearly not as nice as having the parameters passed directly, but it works to allow you to pass multiple parameters and access them using Web API. JObject is JSON.NET's generic object container which sports a nice dynamic interface that allows you to walk through the object's properties using standard 'dot' object syntax. All you have to do is cast the object to dynamic to get access to the property interface of the JSON type. Additionally JObject also allows you to parse JObject instances into strongly typed objects, which enables us here to retrieve the two objects passed as parameters from this jquery code:var album = { AlbumName: "PowerAge", Entered: "1/1/1977" } var user = { Name: "Rick" } var userToken = "sekkritt"; $.ajax( { url: "samples/PostAlbum", type: "POST", contentType: "application/json", data: JSON.stringify({ Album: album, User: user, UserToken: userToken }), success: function (result) { alert(result); } }); Summary ASP.NET Web API brings many new features and many advantages over the older Microsoft AJAX and REST APIs, but realize that some things like passing multiple strongly typed object parameters will work a bit differently. It's not insurmountable, but just knowing what options are available to simulate this behavior is good to know. Now let me say here that it's probably not a good practice to pass a bunch of parameters to an API call. Ideally APIs should be closely factored to accept single parameters or a single content parameter at least along with some identifier parameters that can be passed on the querystring. But saying that doesn't mean that occasionally you don't run into a situation where you have the need to pass several objects to the server and all three of the options I mentioned might have merit in different situations. For now I'm sure the question of how to pass multiple parameters will come up quite a bit from people migrating WCF REST or ASP.NET AJAX code to Web API. At least there are options available to make it work.© Rick Strahl, West Wind Technologies, 2005-2012Posted in Web Api   Tweet !function(d,s,id){var js,fjs=d.getElementsByTagName(s)[0];if(!d.getElementById(id)){js=d.createElement(s);js.id=id;js.src="//platform.twitter.com/widgets.js";fjs.parentNode.insertBefore(js,fjs);}}(document,"script","twitter-wjs"); (function() { var po = document.createElement('script'); po.type = 'text/javascript'; po.async = true; po.src = 'https://apis.google.com/js/plusone.js'; var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(po, s); })();

    Read the article

  • Calling Web Services in classic ASP

    - by cabhilash
      Last day my colleague asked me the provide her a solution to call the Web service from classic ASP. (Yes Classic ASP. still people are using this :D ) We can call web service SOAP toolkit also. But invoking the service using the XMLHTTP object was more easier & fast. To create the Service I used the normal Web Service in .Net 2.0 with [Webmethod] public class WebService1 : System.Web.Services.WebService { [WebMethod] public string HelloWorld(string name){return name + " Pay my dues :) "; // a reminder to pay my consultation fee :D} } In Web.config add the following entry in System.web<webServices><protocols><add name="HttpGet"/><add name="HttpPost"/></protocols></webServices> Alternatively, you can enable these protocols for all Web services on the computer by editing the <protocols> section in Machine.config. The following example enables HTTP GET, HTTP POST, and also SOAP and HTTP POST from localhost: <protocols> <add name="HttpSoap"/> <add name="HttpPost"/> <add name="HttpGet"/> <add name="HttpPostLocalhost"/> <!-- Documentation enables the documentation/test pages --> <add name="Documentation"/> </protocols> By adding these entries I am enabling the HTTPGET & HTTPPOST (After .Net 1.1 by default HTTPGET & HTTPPOST is disabled because of security concerns)The .NET Framework 1.1 defines a new protocol that is named HttpPostLocalhost. By default, this new protocol is enabled. This protocol permits invoking Web services that use HTTP POST requests from applications on the same computer. This is true provided the POST URL uses http://localhost, not http://hostname. This permits Web service developers to use the HTML-based test form to invoke the Web service from the same computer where the Web service resides. Classic ASP Code to call Web service <%Option Explicit Dim objRequest, objXMLDoc, objXmlNode Dim strRet, strError, strNome Dim strName strName= "deepa" Set objRequest = Server.createobject("MSXML2.XMLHTTP") With objRequest .open "GET", "http://localhost:3106/WebService1.asmx/HelloWorld?name=" & strName, False .setRequestHeader "Content-Type", "text/xml" .setRequestHeader "SOAPAction", "http://localhost:3106/WebService1.asmx/HelloWorld" .send End With Set objXMLDoc = Server.createobject("MSXML2.DOMDocument") objXmlDoc.async = false Response.ContentType = "text/xml" Response.Write(objRequest.ResponseText) %> In Line 6 I created an MSXML XMLHTTP object. Line 9 Using the HTTPGET protocol I am openinig connection to WebService Line 10:11 – setting the Header for the service In line 15, I am getting the output from the webservice in XML Doc format & reading the responseText(line 18). In line 9 if you observe I am passing the parameter strName to the Webservice You can pass multiple parameters to the Web service by just like any other QueryString Parameters. In similar fashion you can invoke the Web service using HTTPPost. Only you have to ensure that the form contains all th required parameters for webmethod.  Happy coding !!!!!!!

    Read the article

  • Extracting data from internet

    - by Ankiov Spetsnaz
    I would like to extract data from internet like www.mozenda.com does but I want to write my own program to do that. Specific data I'm looking for is various event data. Based on my research, I think custom web crawler is my answer but I Would like to confirm the answer and see if there are any suggestion to make custom web crawlers if web crawler indeed is an answer. Personally, I would prefer Java and I'm planning on using Glassfish technology if that matters...

    Read the article

< Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >