Search Results

Search found 59112 results on 2365 pages for 'web internet'.

Page 277/2365 | < Previous Page | 273 274 275 276 277 278 279 280 281 282 283 284  | Next Page >

  • How do I configure multiple domain names on my IIS server? [closed]

    - by Dillie-O
    We have a few websites that we are running on one instance of IIS that need to be mapped for each of their domain names. For example. Site A has the domain name coolness.com Site B has the domain name 6to8Weeks.com Site C has the domain name PhatTech.com When I look at the "Web Site Identification" section of the IIS configuration window, I notice that I can specify an IP address and port, but if I click the Advanced button, I can also configure the site based on host header values as well. How do I configure each site in IIS? Ideally I would like them to all be able to listen to port 80, so I don't have weird URLs, but I'm not sure if I do this using headers, IP addresses, both, or something else.

    Read the article

  • How to direct a Network Solutions domain name to an html website hosted on Google Drive? [on hold]

    - by Air Conditioner
    To begin with, I'd wanted to take advantage of HTML, CSS, and so on to build a website that looks and works just as I'd like it to. I took a look around on how I could make that work, and I soon saw a lifehacker article showing that its possible to host website files on google drive. I then made sure that the folder containing the files was shared publicly throughout the web, and I now have a working 'google drive hosted' domain for the website. However, I did want to have the custom domain, and so I registered one with network solutions. So now, I'm curious on how I should direct my Network Solutions domain to the index.html I'm hosting on google drive. Would anyone have an Idea?

    Read the article

  • Can't open Control Panel or IE

    - by Josh
    I have a XP computer where when ever I try to open Control Panel nothing happens, nothing flashes on the screen. Same thing with Internet Explorer. I've scanned the computer with Malwarebytes and Avast, Malwaresbytes found some Adware which it removed without problems. Avast found nothing. I looked at the running processes with Process Explorer, nothing malicious running. And looked at a Process Monitor output when I tried to run IE, nothing obviously wrong. The process just decides to exit. What can I try next? I would suspect corrupt IE install but Control Panel doesn't work either. UPDATE: Nether work in Safe Mode under the user account. (only 1 user on the computer) But in Safe Mode, under the built in Administrator account, they work. So what ever is broken, is only broken in the one account. Anything under the HKCU registry key that could break this?

    Read the article

  • Visual Web Part as a Sandboxed solution

    - by Steve Clements
    You want the RAD wonderfulness of a visual web part, but it needs to be deployed as a Sandboxed solution. Problem? No, SharePoint powertools for visual studio to the rescue!   http://goo.gl/pQ9ct   There are a couple limitations, read the above page, nothing major. e.g. 1. Javascript debugging is not supported 2. Debugging asp.net code is not supported. 3. Use of <% Assembly Src= is not supported   I understand it does it by adding the markup as an embedded resource, but I haven't actually tried it yet!  To come!

    Read the article

  • Tell a user whether they have already viewed an item in a list. How?

    - by user2738308
    It is pretty common for a web application to display a list of items and for each item in the list to indicate to the current user whether they have already viewed the associated item. An approach that I have taken in the past is to store HasViewed objects that contain the Id of a viewed item and the Id of the User who has viewed that item. When it comes time to display a list of items this requires querying for the items, and separately querying for the HasViewed objects, and then combining the results into a set of objects constructed solely for the purpose of displaying them in the view. Each e.g li then uses the e.g. has_viewed property of the objects constructed above. I would like to know whether others take a different approach and can recommend alternative ways to achieve this functionality.

    Read the article

  • How do non-coders do simple local templating to avoid redundant HTML? [closed]

    - by Max Cantor
    I'm a web developer. When I start designing a site, I use a framework to handle templating for me, even if it's just rack + erubis. What do non-developers do? If you want to implement a site in HTML and CSS without a framework running on a webserver, without frames, and without WYSYWIG tools like Dreamweaver... how do you avoid copy-and-pasting the HTML of your navigation (for example) on every single page you're writing? I feel stupid asking this because it seems like their must be an obvious answer, but for the life of me, I can't think of one right now.

    Read the article

  • Implementing Command Pattern in Web Application

    - by KingOfHypocrites
    I'm looking to implement the command pattern in a web application (asp.net c#)... Since the commands come in text format from the client, what is the best way to translate the string to a command object? Should I use reflection? Currently I just assume the command that comes in matches the file name of a user control. This is a bit of a hack. Rather than have a select case statement that says if string = "Dashboard" then call Dashboard.Execute(), is there a pattern for working with commands that originate as strings?

    Read the article

  • Removing unnecessary joomla CMS components

    - by Gaz_Edge
    Background I have been developing a web application. I have been using Joomla platform and CMS as the base from my application. My application relies heavily on the joomla platform, but uses very few features of the CMS (I have to login users etc so I use the _users component of the CMS). Question What is the best way to start removing some of the excess CMS that I don't need? For example when I being building my production site, I will want to avoid creating a load of joomla component database tables etc. I have read that an application can be build from the ground up using just the joomla platform, but i still want to include the _users component and would also like to avoid having to start coding all the session handling aspects. Any advice?

    Read the article

  • Why my web site was visited by ARPA?

    - by ilhan
    20 minutes ago a user-agent with 66.116.153.122 IP address has visited my web site. It's domain is rev.opentransfer.com.122.153.116.66.in-addr.arpa. User-agent: Mozilla/5.0 (Windows; U; Windows NT 6.1; de; rv:1.9.2.11) Gecko/20101012 AskTbSPC2/3.8.0.12304 Firefox/3.6.11. Language: en-us,en;q=0.5,de-de,de;q=0.8. Compression: gzip,deflate. Oh, and my domain name ends with .name. Why ARPA has visited me?

    Read the article

  • Web pages with mixed ownership photos

    - by dstonek
    I have a photo website. 15% of the photos belong to approved registered users. They agree my terms about uploading their images in my web pages. I include a photographer credit on right bottom corner. About identifying the site with google, every page contains a google+ button to MY google+ page it also contains <link href="https://plus.google.com/nnnnnnnnnn/" rel="publisher" /> I need some advice in order to respect google rules about my pages containing other photographers images not to be penalized because of possible duplicated or interpreted as stolen content. My concern is also about adding G+ links (to MY photo page) and Google publisher id would harm my site rank because of pages containing third-party photos.

    Read the article

  • questions on transfer rate

    - by Nrew
    I tested my internet speed at speedtest.net. And go the result. .42Mbps Download and .21 Upload rate. My classmate got the same download speed of .42Mbps but has .87Mbps upload rate. Does upload rate affect the transfer rate?Because even though we got the same download speed. His transfer rate is about 100kbps downloading a movie from a torrent. And mine is only about 47kbps. Also the same torrent. And even direct download its always 47kbps. Is it possible to tweak something in order to have higher transfer rates.

    Read the article

  • English major new to programming. What language should I learn first? [closed]

    - by PJKaka
    After working extensively an internet startup in a marketing positions, I've decided to wade into the entrepreneurship pool with a startup of my own. The only problem: I don't have any particular technical skills to speak of. Although I can find a technical co-founder, I'd rather not be the stereotypical 'business guy' drumming his fingers on the desk and asking 'how much longer?' as my technical co-founder codes away. I would like to understand code and what's happening in the backend, even if I don't end up being anything more than a 'passable' programmer. With this in mind, which language should I try to learn first? For the record, I'm quite proficient with HTML, CSS, and a bit of JavaScript. I have some familiarity with PHP because I've toyed around with WordPress a lot, but my knowledge is limited at best. My math skills are quite strong. I took some advanced calculus courses in college since I've always enjoyed the subject. While my goals are to learn web development, I wouldn't mind learning some hardcore object oriented programming skills in C or Java as well.

    Read the article

  • why only google related sites load using wireless network

    - by gansai
    I have a wireless internet connection through BSNL. I have a windows 7 ultimate, and using latest Google chrome. During night time ie from 12:30 am onwards till around 5:30 am, I have the following problem: Only the following webpages load: Google Search Google images and all related to google product pages like you see them on the google home page. Blogspot webpages as supported through Google. Any other site loading gets stuck with the status message:- Waiting for www.thaturl.com .... And it goes on and on. Why does this happen? I checked around googling for this but no solution at all. By the way I tried changing my dns addresses too sometimes from automatically to opendns and then Google public dns. I need some solution which might help me to load other websites also. Thanks in advance.

    Read the article

  • What music streaming app fits my needs on Ubuntu?

    - by Jim
    I'm looking for an application to stream Internet radio on Ubuntu. I like listening to Radio Paradise while I work. Right now, I'm using Amarok. "Movie Player" sometimes refuses to open the stream, and VLC doesn't keep its window title updated with the currently playing track. Amarok has nice translucent notifications when tracks change, but track changes in streams don't trigger the notifications. Mostly, I want something that reliably opens streams and makes it easy to see the name of the track that's playing. If it has a built-in directory of streaming radio stations, that would be a big benefit.

    Read the article

  • What are good/fast methods to pull data from a database using JavaScript?

    - by Yatrix
    I'm pretty new to web technologies and I am creating a filter control that will have cascading controls. We are doing a lot of this through JavaScript and are debating the best route to take to the database. HTTPHandlers, WebServices and Ajax are all being considered (or a combination of them). We want to be able to handle a million rows in theory, so it has to be scalable to at that. We are going through JavaScript as our page must not do post-backs, if your'e wondering. I'm asking from an architectural standpoint, but will take any useful information. Links, control suggestions - anything you have, I'll happily listen to.

    Read the article

  • robots.txt dissalow url containing string with a '/' at the end

    - by thanili
    i have a website with thousands of dynamic pages. I want to use the robots.txt file in order to dissalow certain url patterns corresponding to pages with duplicate content. For example i have a page for article itemA belonging to category catA/subcatA, with URL: /catA/subcatA/itemA this is the URL that i want to be indexed from google. this article is also visible via tagging in various other places in the web site. The URLs produced via tagging is like: /tagA1/itemA this URL i want NOT to be indexed from google. However i want to have indexed all tag listings: /tagA1 so how can i achieve this? dissalow URLs of including a specific string with a '/' at the end? /tagA1/ itemA - dissalow /tagA1 - allow

    Read the article

  • How long till HTML5 canvas becomes a viable game development platform?

    - by Shouvik
    So I have been working on web application. So invariably what it boils down to is making simple games which were previously based on flash or openGL. Now I know apple was moving away from flash because its proprietary unlike their stance that its got "pathetic performance"! Not true, try playing a canvas game, I can assure you at any point of time (including when its idle) it will use up a fair bit of processing power just to redraw the UI. Now I do understand that this is my fault because when the game is not active I should not be redrawing the canvas, but honestly its a lot of work and I suppose there should be libraries which should be able to assist me with that! So, how much will it be before I see a decent canvas library which handles these "tiny" issues for me? I can't honestly expect Steve Jobs to be doing anything more for HTML5! I someone knows of a good library, I am all ears...! :) PS: I use mootools and am presently using Mootools Canvas Library.

    Read the article

  • Suggest windows webhost provider for following requirements.

    - by op_amp
    Hi, We have a asp.net MVC3 based web app which uses SQL SERVER 2008 for database. Also, we have a client side desktop application which also uses SQL SERVER 2008. While developing the system, we are able to Sync tables using SQL SERVER Replication feature. Now, we want to host our site on a webserver but we are clueless about it. If anyone of you have a similar system working then please suggest a cheap but reliable webhost which supports Replication. Initially there will be approximately 10 or less clients who will perform replication 2 or 3 times a day. The size of the database will be less than 4GB for sure.

    Read the article

  • Image caching when rendering the same images on different pages

    - by HelpNeeder
    I'm told to think about caching of images that will be displayed on the page. The images will be repeated throughout the website on different pages and I'm told to figure out the best way to cache these images. There could be few to dozen of images on single page. Here's few questions: Will browser caching work to display the same images across different web pages? Should I rather store images in stringified form in a memory instead, using JavaScript arrays? Store them on hard drive using 'localStorage'? What would be easiest yet best option for this? Are there any other alternatives? How to force cache? Any other information would be greatly appreciated...

    Read the article

  • Password protected traffic meter

    - by UncleBob
    Hi first, I have a small problem for which I haven't found a solution yet. I live in Bosnia and share the Internet connection with the landlady, and as is usual in Bosnia, we do not have a flat rate, but a 15 Giga traffic limite. That would actually be more than enough, if the son of the landlady wouldn't be watching videos all the time, so the bills are truning out rather expensive. I have already installed a traffic monitoring program, but he apparently turns it off as soon as he comes close to his limit and then denies that he consumed any more. I therefore need at least a measurement program that is password protected and / or notes in the log when it's been turned off. Even better would be a program that just cuts his access when he exceeds his share, ie a mixture of Traffic meter and Parental Guard. Can someone help me out here?

    Read the article

  • Navigation for ASP.NET Web Forms project published on codeplex

    Navigation for ASP.NET Web Forms manages movement and data passing between aspx Pages in a unit testable manner. There is no Client-side logic, so it works in all browsers, and no Server-side cache, so it works with the browser back button.Features include loosely coupled Pages, typed data passing, empty code-behinds, context-sensitive bread crumb trail, ASP.NET Data binding integration, automatic ASP.NET Ajax history navigation and many more.The source code, binaries and comprehensive documentation...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • How can I restrict my mates to stop downloading?

    - by user239295
    We are sharing an internet broadband connection with 6 users at a place we live. We get 20 gb fup ( Fair usage policy) with 2 mbps speed from the ISP after the 20 gb is consumed the speed comes down to 512 kbps very difficult to browse any page. The problem is we cannot track which user/mate is downloading and ending the FUP. it is very difficult to track so is there something that we can allot per user some amount of space like 2 gb of downloading or restrict all from downloading so that we can utilize all the fup till the end of the month. We are using this connection as wifi configured. A adsl router is configured as wifi and we all using all 6 laptops. No PC. Any help would be appreciated. I apologize if i am not clear with my question.

    Read the article

  • HTML, JS, CSS Engines

    - by Pius
    I am just messing around, trying to figure out how stuff works and right now I have a couple questions about HTML, JS and CSS engines. I know there are two major JavaScript engines out there - V8 and JavaScriptCore (WebKit's JS engine as far as I know). Is that correct? And what are the main HTML + CSS renderers out there? Let's say I want to build a web browser using V8 (I saw it has some documentation and stuff + I like the way it works), what are the best options for me? Partially another question. Is there any bare browser that uses V8 and runs on Ubuntu at least? P.S. I am a Ubuntu user and prefer C++.

    Read the article

  • Switching to Chrome from IE

    - by Alan Parrish
    Hi, I work in IT at a school and we recently updated our database software, however the web access that the teachers use to do registration does not work too well with Internet Explorer 7 and we're thinking about switching to chrome (mostly due to me hating IE in general) but the problem we have is that users are unable to migrate their bookmarks over from IE due to account restrictions, is there any way to get this working? A bit of info about the system if it could be helpful; Almost all the client PCs are running Win XP SP2. My colleagues machines are running Windows 7, I use OS X Snow Leopard and most of out servers are running windows server 2003 (with the exception of 2 one on 2000 server and another on 2008 R2). The Active Directory Controller is running Windows Server 2003

    Read the article

  • Getting started with SOAP [closed]

    - by EmmyS
    A site I developed has a new requirement to get weather data from the National Weather Service. They have quite a bit of info on how to use SOAP to get their data and display it in the browser, but what we need to do is use a cron job to get the data at specific intervals, then parse the data out into a database. I have no problem writing PHP code that will run an XSLt and parse xml records out into SQL queries, but I have no idea how to handle this with SOAP (which I've never worked with.) Do I get the data via a SOAP request, save it to an XML file on my web server, then run the XSLt against that? Or is there some other way to go about this?

    Read the article

< Previous Page | 273 274 275 276 277 278 279 280 281 282 283 284  | Next Page >