Search Results

Search found 14372 results on 575 pages for 'home premium'.

Page 175/575 | < Previous Page | 171 172 173 174 175 176 177 178 179 180 181 182  | Next Page >

  • Highlight current page tab [migrated]

    - by Jose David Garcia Llanos
    I am making a website, currently i am setting the highlight tab for current page, a particular page is not highlighting its tab and i have checked the code about 5 times but i cant find anything wrong with it. the website is auto-sal.es Here is the code: style.css body#home a.hometab, body#cars a.cartab, body#feedback a.feedtab, body#contact a.contacttab, body#members a.memberstab {background: #7D0000;} contactus.html <body id="contact"> navigation <ul id="menu"> <li><a href="index.html" target="_self" class="hometab">Home</a></li> <li><a href="cars.html" target="_self" class="cartab">Cars</a></li> <li><a href="feedback.html" target="_self" class="feedtab">Feedback</a></li> <li><a href="contactus.html" target="_self" class="cotacttab">Contact Us</a></li> <li><a href="members.html" target="_self" class="memberstab">Members</a></li> </ul> Again, the issue is that it is not highlighting the tab for contact us

    Read the article

  • capture login screen-not sucssecful

    - by yinon
    I'm using Ubuntu 12.04 . If more information is needed, please tell me. I Tried some guides to cature my login screen: How can I take a screenshot of the login screen? http://www.howtoforge.com/how-to-take-a-screenshot-of-your-login-screen I tried the first one, but after running: $ sudo bash /tmp/shot.sh >/tmp/shot.xwd I'm getting this: No protocol specified No protocol specified xwd unable to open display ':0' and i'm getting an xwd file in my tmp folder. running the other commands giving errors. tried the second one, after running: chvt 8; sleep 5; XAUTHORITY=/var/gdm/:0.Xauth DISPLAY=:0.0 import -window root /tmp/gdm-login-shot.png I got: No protocol specified No protocol specified import: unable to open X server ':0.0' @ error/import.c/ImportImageCommand/366 Many Thanks for helping (: EDIT to vine_user: here the output from the terminal-I cacked it from within the system while I'm looged in: ubuntu@PrecisePangolin:~$ echo 'sleep 5; DISPLAY=:0 XAUTHORITY=/var/run/lightdm/root/$DISPLAYDISPLAY xwd -root' > /tmp/shot.sh ubuntu@PrecisePangolin:~$ sudo bash /tmp/shot.sh >/tmp/shot.xwd bash: /tmp/shot.xwd: Permission denied ubuntu@PrecisePangolin:~$ sudo su root@PrecisePangolin:/home/ubuntu# sudo bash /tmp/shot.sh >/tmp/shot.xwd No protocol specified No protocol specified xwd: unable to open display ':0' root@PrecisePangolin:/home/ubuntu# EDIT 2!: HERE IT IS!: just uploading-this uploadong site is better than other iuset to use (: :

    Read the article

  • Setting up nginx as proxy to apache; All good, but nginx doesn't serve media

    - by becomingGuru
    I have set it up such that nginx proxies request and sends django requests to apache and serves media itself. Following documents my setup: Nginx Configuration: /etc/nginx/nginx.conf user www-data; worker_processes 1; error_log /var/log/nginx/error.log; pid /var/run/nginx.pid; events { worker_connections 1024; } http { include /etc/nginx/mime.types; default_type application/octet-stream; access_log /var/log/nginx/access.log; sendfile on; tcp_nopush on; #keepalive_timeout 0; keepalive_timeout 65; tcp_nodelay on; gzip on; include /etc/nginx/sites-enabled/*; } ===== ngnix proxy /etc/nginx/proxy.conf ============ proxy_redirect off; proxy_set_header Host $host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; client_max_body_size 10m; client_body_buffer_size 128k; proxy_connect_timeout 90; proxy_send_timeout 90; proxy_read_timeout 90; proxy_buffer_size 4k; proxy_buffers 4 32k; proxy_busy_buffers_size 64k; proxy_temp_file_write_size 64k; =========== Nginx server file: /etc/nginx/sites-enabled/some-name.txt ========== server { listen 208.109.252.110:80; server_name netconf; autoindex on; access_log /home/site/server_logs/nginx_access.log; error_log /home/site/server_logs/nginx_error.log; location / { proxy_pass http://127.0.0.1:80/; include /etc/nginx/proxy.conf; } location /site_media/ { root /home/site/folder/static; } } ========== Nginx very well proxies the request and passes to apache, the required requests, but doesn't serve the media. In the last server file, location site_media is not served, at all. :( Everything seems perfect to me. What is wrong? Thanks in advance.

    Read the article

  • Can .htaccess slow down a site?

    - by Cody Sharp
    I'm working with a client on an e-commerce website. I implemented clean URLs using .htaccess. I also used .htaccess to solve canonical issues such as redirecting www to non-www and removing index.php from the URL. The website recently began to slow down dramatically, sometimes not even loading. The site is hosted on GoDaddy, and when the client called GoDaddy they told him it was the .htaccess file slowing down the website. I find this highly unlikely because of my past experiences, but I'm not 100% sure. My thinking is that the client's website is most likely on a shared server with a busy neighborhood, thus slowing down the site. It's not always slow, but rather sporadic throughout the day, loading fast at some points and slow at other points in time. Can the .htaccess file slow down a website to a crawl? If so, are there better ways to solve these problems with different rewrite rules and such? Here is what the actual .htaccess file looks like: Options +FollowSymlinks RewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} ^www.example.net [NC] RewriteRule ^(.*)$ http://example.net/$1 [L,R=301] RewriteRule ^products/([0-9a-zA-Z\_\-]*)\.htm([l]?)$ index.php p=product&product_code=$1 [L] RewriteRule ^catalog/([0-9a-zA-Z\_\-]*)\.htm([l]?)$ index.php p=catalog&catalog_code=$1 [L] RewriteRule ^pages/([0-9a-zA-Z\_\-]*)\.htm([l]?)$ index.php?p=page&page_id=$1 [L] RewriteRule ^index\.htm([l]?)$ index.php?p=home [L] RewriteRule ^site_map\.htm([l]?)$ index.php?p=site_map [L] RewriteCond %{QUERY_STRING} ^p=home$ RewriteRule (.*) ? [R=permanent] I'm a .htaccess and regex novice, so any pointed out mistakes would also help. Thank you.

    Read the article

  • File/folder permissions and groups on Linux with Apache

    - by phobia
    I'm trying to learn about permissions on linux webserver with apache. Some clues to the system: The server I have to play around with is Fedora based. Apache runs as apache:apache. To allow for e.g. php to write to a file the file needs to be chmod 777. 755 is not sufficiant. What I'm wondering is basically how set up permissions like they should be on e.g. a "shared web host". My main problem is that if I set a permission so that one user cannot access anothers home folder, then apache can't read from the public_html folder either. To keep the users out I need to set chmod 700. But to let apache to read I need to have at least execute on world, so a 701 basically works, but won't let some users in. So I'm really stuck on what to do. Have been concidering adding the apache user to the frous grours below to avoid having to add the world execute flag, but is that a bad thing? Should it be the other way around, the users in the groups below should also be in the apache group? I was aiming at having 4 groups: 1. webapp same as dev_int, but is the only one that can go inside the webapp/live folder to e.g. do an update from the repo. 2. dev_int can read,write and execute everything in the "web root", including the two below, but nothing outside of the web root 3. dev_ext can read write and execute in all client folders, but cannot access anything outside of the webapp root 4. clientsBasic ftp accounts. Has a home folder with a public_html, but cannot access any other home folders An example of folder structure: webroot    no users in the aforementioned groups can go outside of here some_project    :dev_int only webapp live    :webapp only staging    :dev_int and :dev_ext clients    :dev_int and :dev_ext client_1    :dev_int, :dev_ext and client1:clients public_html dev developer_1    developer_1:dev_int OR :dev_ext public_html

    Read the article

  • Boot linux off hard drive and then switch to run from usb flash disk

    - by Jesse
    I have an older laptop that I want to use as a simple media server on my home network. I would like to avoid using the internal hard drive except for booting (BIOS does NOT support booting from USB). My thought was to mirror the hard drive (currently has current install of Arch Linux) onto the flash drive and then after booting switch over to run everything from the flash drive. I read the following article about using a RAM disk (HOW-TO: Boot OS into RAM for speed and silence) but ran into problem because the USB subsystem does not seem to be initialized soon enough (I create root and home paritions on the flash disk and modified fstab to pick those - didn't work). Any thoughts?

    Read the article

  • Dualboot 12.04/windows 7 After installation from USB reboot straight to Windows with no option to select ubuntu and no boot loader

    - by Alkatraz
    windows 7 home premium intel i5 2500k CPU ASUSP8Z68-V PRO Motherboard GeForce GTX 570 GPU corsair 120Gb SSD (windows 7 os) WD 1tb HDD I select the USB drive in the BIOS and boot to it and choose install. i select to manually configure partitions, partition the 200gb of unallocated space on my 1Tb HDD into 16Gb swap file 30Gb / extf4 and 154Gb /home extf4. I make sure that the boot loader is installed to corsair 120Gb SSD (where the windows boot is) and installation goes smoothly. When i reboot after install it runs through bios strait into windows. I have tried upwards of a dozen times and i have also tried with linuxmint. I have also redownloaded the ISO and used two different programs to create the live usb. the installation seems to go well as i can see the partitions i have created in the windows device manager after install http://imgur.com/Wp0V1 I currently run lubuntu on my laptop but it is not a dual boot. i'm assuming this is a boot loader issue and i am assuming that inside those partitioned files in my screenshot there is a working OS of ubuntu 12.04 i just have no way of getting to it.

    Read the article

  • Why is my system waiting for google-analytics?

    - by mikeroosa
    When I go to certain websites that have google analytics, they never load up. One example is www.jquery.com. I tried in firefox, chrome, safari and explorer. Chrome & safari, flash the home page then it disappears and the screen goes white. Firefox never loads anything and the spinner just keeps spinning. Loads fine in explorer. Any ideas of what may be going on? If I use my laptop, on my same home internet connection, it works fine. This is not happening on all sites with analytics, just certain ones.

    Read the article

  • Why is my system waiting for google-analytics?

    - by mikeroosa
    When I go to certain websites that have google analytics, they never load up. One example is www.jquery.com. I tried in firefox, chrome, safari and explorer. Chrome & safari, flash the home page then it disappears and the screen goes white. Firefox never loads anything and the spinner just keeps spinning. Loads fine in explorer. Any ideas of what may be going on? If I use my laptop, on my same home internet connection, it works fine. This is not happening on all sites with analytics, just certain ones.

    Read the article

  • What is Ubuntu's Definition of a "Registered Application"?

    - by Tom
    I've run into this a few times when installing apps from source, and during the occasional hack with update-alternatives. So far, it's only been a minor annoyance (ie, not got in the way of the end-goal) but it's now a frustration as it's pointing to a hole in my knowledge-base... so when I get a message that 'foo' is "not a registered application" (or I can't use foo's default icon cuz Ubuntu has no knowledge of 'foo'): (1) what defines a "registered application"? (2) how can I define an application installed from source (and likely residing in $HOME/bin/app-name) such that it packs the same functionality as a package installed from a .deb? (if the solution is not self-evident from answer 1) Example: I download and unpack daily dev builds of sublime-text-2 to /home/tom/bin/sublime-text-2. I've created a *.desktop file with appropriate shortcuts, etc. But the icon for sublime cannot be display in any launcher even if I provide a full pathname to the option. The solution is to install a 2nd instance of sublime from a deb package. When I install sublime-text-2 from a .deb package, it installs under /usr/bin && /usr/lib, the installed .desktop file is stored under /usr/share/applications, and the relevant line reads: icon=sublime_text. Where's the linkage I'm missing? Somehow Ubuntu knows how to exact the icon from sublime_text in the latter, but not in the former (again, even with a full path provided).

    Read the article

  • Share files from Windows XP to Mac Snow Leopard

    - by sympleko
    Hi, I have a Windows XP desktop and a MacBook Pro on my home network. I would like to mount my "home directory" (My Documents, or whatever they call it in Windows) onto my MacBook's filesystem. So far I have been able to do this with the Shared Documents folder, using the excellent howto at Maclive. But I'd like to be able to authenticate using my Windows XP username and password, and access my files remotely without exposing them to everyone on the nextwork. Any clues or good links?

    Read the article

  • Spotlight on an Office – Reading TVP offices

    - by Maria Sandu
    This month we’re in the UK at the Reading offices, for ‘Spotlight on an office’. The Reading Office, which is Oracle’s UK Headquarters, is based in Thames Valley Park (TVP), which is a bustling hive of activity that houses many different companies, a gym, and even a nursery. Overlooking the Thames and some of England’s beautiful countryside, this office, just a short free bus ride from Reading Town Centre is in a fantastic location. The offices themselves are made up of 5 different buildings, each with their own car park, restaurant, and design. The main building or TVP 510 as it is referred to, sits resplendent next to an extremely blue (for the UK) pond, filled with large koi-carp that on a sunny day like to come to the surface of the lake and bask. As the main hub of activity, TVP 510 is where you will find our Dry Cleaning service, the Ozone Gym, the main restaurant (which never fails to have someone in it), and the Marquee which sits outside the back amongst the picnic benches, and is where we have Barbeques in the summer time. Another highlight of the Reading Offices is tucked away in TVP530; the home of H20, and our sports and social club. This is the building that can be best be described as having the ‘cool’ vibe, where you can relax and unwind, all whilst sipping a Starbucks (or Costa if you prefer, located in TVP550), and playing a game of Pool in the cafeteria, or alternatively you can sit back and enjoy a seat in one of the luxury massage chairs! If you feel so inclined, you can also hire out an OraBike from any of the TVP offices, and if you are anything like some of my team, cycle from Reading to Bath using the towpath starting in Thames Valley Park. Oracle’s Reading Offices are a great place to work, they are home to a diverse range of people and have great atmosphere which would suit a graduate, intern, or anyone who is looking to come and work for Oracle in the UK.

    Read the article

  • remote desktop over wireless

    - by tbischel
    So I'm trying to run remote desktop on my laptop to connect to my home desktop. I have a problem where this works fine if I connect my laptop with an ethernet cable, but fails when I try to use wireless internet access (which works fine for normal internet surfing). I've experienced this problem at home with my wireless router, and at work with the wireless network they have there, so I'm inclined to believe that its a setting local to my machine rather than the router blocking the requests... but I'm not sure where to look. Any suggestions?

    Read the article

  • Windows Firewall: How to allow traffic on port 8080?

    - by Chadworthington
    I am trying to configure team Foundation Server so that 1) it is accessible from within my Home Network 2) and then make the Web site access accessible via the Internet I have a problem with point 1: When I access http://192.168.1.106:8080/tfs/web/ locally from 192.168.1.106, it works. When I access the same web site from another PC in my home network, the abive URL works only if I turn of the Firewall on 192.168.1.106. Can someone please tell me specifically how to allow traffic on port 8080 without turning off Windows Firewall? It seems that the exceptions that I specify are intended for listing programs on the box that need to communicate out. Is IIS the program that I need to make the exception? How do I specify that port 8080 traffic should be allowed for web site traffic on this port? I hope to have success with pt. 2 later but I figure (1) should be done first. I expect issues.

    Read the article

  • Installing Gitolite on NAS - FATAL: have errors but logging failed

    - by Jay
    I'm trying to install Gitolite on my Synology DiskStation, following these instructions: http://www.bluevariant.com/2012/05/comprehensive-guide-git-gitolite-synology-diskstation/ Under "Install Gitolite on the DiskStation and Run Setup" When I run the gitolite install command: DiskStation> /volume1/homes/git/gitolite/install -ln /bin I get the following error: FATAL: have errors but logging failed! 2012-05-31.00:10:22 no GL_LOGFILE env var 2012-05-31.00:10:22 die could not symlink /volume1/home/git/gitolite/src/gitolite to /bin<<newline>> at /volume1/home/git/gitolite/install line 71<<newline>> I am very new to all of this. Does anyone know what this means, and how I can fix this error? Thank you very much.

    Read the article

  • How to stop Windows 7 from automatically connecting to unsecure wifi network

    - by Remi Despres-Smyth
    One of my neighbors has an unsecure wifi network called WLAN. At one point in the past, I accidentally connected to it, and disconnected immediately when I noticed. Now, when I open my laptop at home, it sometimes connects to the WLAN network first, before trying my (secured) home wifi network. The information I've found regarding this issue seems to suggest this network should have a profile on the "Manage wireless networks" screen - but it does not. How do I tell Windows 7 to never connect to networks with SSIDs called WLAN? Or to never connect to unsecured networks without confirming with me first?

    Read the article

  • FTP Reporting Disk Quota Exceeded

    - by Austin
    I am using Notepad++ with FTP_Synchronize to upload files to a server, however, it appears that it is not allowing my file to upload because apparently the "Disk Quota Exceeded" 11:18:49 > -> TYPE I 11:18:49 > Response (200): Type set to I 11:18:49 > -> PASV 11:18:49 > Response (227): Entering Passive Mode (*,*,*,*,*,*). 11:18:50 > -> STOR /home/*/../../var/www/html/test.html 11:18:50 > Response (150): Opening BINARY mode data connection for /home/*/../../var/www/html/test.html 11:18:50 > Response (552): Transfer aborted. Disk quota exceeded Now it may appear that yeah my Disk quota is exceeded, however I've gone to the back-end and saw: Total Used Bandwidth 107.055 MB Allowed Quota 3,000.0 MB Note: Stars were put in place for irrelevant data.

    Read the article

  • My DNS cannot resolve an web site address?

    - by ipkiss
    Hello all, Recently, I could not access the webpage bbc.co.uk anymore, while I can access other websites smoothly. Ar first, I though there may be some problem with my laptop. However, if I use my laptop through my company network, I can load the page bbc.co.uk normally. Then, I though maybe my ADSL at home blocks that web address. However, I tried another laptop with my home ADSL and it can load the page bbc.co.uk very fast. Now I do not know what could be the problem. Can anyone tell me please? Thank you.

    Read the article

  • Google local search rankings is it possible without the use of citations

    - by bybe
    I have a client that is wanting a website design for his self-run business... Basically he is a self employed plumber so his home address is not visitable by the public, however the problem here is that he does not want his home address visible on the internet at all for one reason or another. I have informed him the benefits of having his address visible for such reasons as trust by customers as well as the benefits via Google's local search algorithms (Citations - Visible Address Details) on various directories including Google Maps, and Google Places. But he is clear that he does not want his address online and wants SEO + Web Design without any citations. Now, I care about my reputation in my local area and do not like do half-cut jobs, If I do SEO I want them to be the best they can otherwise word of mouth that customer could say to someone else after my services they are no where to be seen, (I know you can't keep them all happy but none the less). This is kinda new for me since Google introduced local rankings and something I've never had to do... So my question is fairly simple and hope that people who reply have some kind of experience in attempting ranking websites locally without citations.. Is it even possible to rank a local website with Google's local algorithms without the use of citations (address information)?

    Read the article

  • Load login shell inside user cronjob

    - by sa125
    I'm trying to run a rake task using a scheduled cronjob. My crontab looks something like this: 0 1 * * 1-7 /bin/bash -l -c "cd ~/jobs/rake && rake reports:create >> ~/jobs/logs/cron.log" Ruby on my account is provided by RVM, which is loaded via ~/.bashrc (before the no-interaction check): # load RVM env [[ -s $HOME/.rvm/scripts/rvm ]] && source $HOME/.rvm/scripts/rvm # If not running interactively, don't do anything [ -z "$PS1" ] && return # ... rest of logic Time and again, this task fails to run since RVM isn't loaded when the task is called (uses system's /usr/bin/ruby instead), and gem dependencies are missing. How can I make crontab load my shell environment before executing my scheduled jobs? thanks.

    Read the article

  • Do you tend to write your own name or your company name in your code?

    - by Connell Watkins
    I've been working on various projects at home and at work, and over the years I've developed two main APIs that I use in almost all AJAX based websites. I've compiled both of these into DLLs and called the namespaces Connell.Database and Connell.Json. My boss recently saw these namespaces in a software documentation for a project for the company and said I shouldn't be using my own name in the code. (But it's my code!) One thing to bear in mind is that we're not a software company. We're an IT support company, and I'm the only full-time software developer here, so there's not really any procedures on how we should write software in the company. Another thing to bear in mind is that I do intend on one day releasing these DLLs as open-source projects. How do other developers group their namespaces within their company? Does anyone use the same class libraries in personal and in work projects? Also does this work the other way round? If I write a class library entirely at work, who owns that code? If I've seen the library through from start to finish, designed it and programmed it. Can I use that for another project at home? Thanks, Update I've spoken to my boss about this issue and he agrees that they're my objects and he's fine for me to open-source them. Before this conversation I started changing the objects anyway, which was actually quite productive and the code now suits this specific project more-so than it did previously. But thank you to everyone involved for a very interesting debate. I hope all this text isn't wasted and someone learns from it. I certainly did. Cheers,

    Read the article

  • PureFTPd : ChrootEveryone not working

    - by Mistha Noobstha
    I have a fresh install of Ubuntu 12.04 with pure-ftpd version 1.0.35-1. I don't use database for the only user I have (ftpuser1 - localuser), but I want to lock him into his home directory. /ftp/ftpuser1 is the home dir of ftpuser1. /etc/pure-ftpd/conf/ChrootEveryone has the yes string and in /etc/default/pure-ftpd-common VIRTUALCHROOT is set to false, but ftpuser1 can freely walk around. My startup line for pure-ftpd looks like this : /usr/sbin/pure-ftpd -l pam -A -8 UTF-8 -O clf:/var/log/pure-ftpd/transfer.log -u 1000 -E -B So there is the "-A" in it for the ChrootEveryone.

    Read the article

  • VPN and Bonjour conflicting

    - by JW.
    Does anyone know why a VPN connection might interfere with Apple's Bonjour? I've noticed that my Mac and various iDevices have trouble finding each other on my local network, when I have VPN connections open. Things like Home Sharing and Wi-Fi Sync work some of the time, but sometimes fail to find the other device. The VPN connections are made using IPSecuritas, which is a GUI around raccoon. I have the local "endpoint mode" set to Host. Apple mentions that Home Sharing may conflict with VPNs, but they don't specify why, or how to fix it. I'm using a Mac with OS 10.7.3 and IPSecuritas to connect to the VPN, an iPhone, and an iPad.

    Read the article

  • trying to install Mumble on centos

    - by ErocM
    I am trying to install mumble on my CentOS vps following these directions: http://www.hosting.com/support/linux/install-the-mumble-voip-server-to-redhat-or-centos When I get to this line: rpm2cpio mumble-server-1.2.2-3mdv2011.0.x86_64.rpm > file.lzma lzma -d file.lzma I am getting this error: root@vps-1112788-12524 [/home/~~~~/mumble]# rpm2cpio mumble-1.2.4-0.20120422.1-mdv2012.0.x86_64.rpm > file.lzma root@vps-1112788-12524 [/home/~~~~/mumble]# lzma -d file.lzma lzma: file.lzma: File format not recognized I did update the name of the file since the link they gave me was not available. I got the new rpm from the same place: http://www.rpmfind.net/linux/rpm2html/search.php?query=mumble+server&submit=Search+... I'm new to Linux so I have no idea what I'm doing wrong. Can anyone help me out?

    Read the article

  • Register now for the UK Windows Azure Self-paced Interactive Learning Course starting May 10th

    - by Eric Nelson
    [Suggested twitter tag #selfpacedazure] We (myself and David Gristwood) have been working in the UK to create a fantastic opportunity to get yourself up to speed on the Windows Azure Platform over a 6 week period starting May 10th – without ever needing to leave the comfort of your home/office.  The course is derived from the internal training Microsoft gives on Azure which is both fun and challenging in equal parts – and we felt was just too good to keep to ourselves! We will be releasing more details nearer the date but hopefully the following is enough to convince you to register and … recommend it to a colleague or three :-) What we have produced is the “Microsoft Azure Self-paced Learning Course”. This is a free, interactive, self-paced, technical training course covering the Windows Azure platform – Windows Azure, SQL Azure and the Azure AppFabric. The course takes place over a six week period finishing on June 18th. During the course you will work from your own home or workplace, and get involved via interactive Live Meetings session, watch on-line videos, work through hands-on labs and research and complete weekly coursework assignments. The mentors and other attendees on the course will help you in your research and learning, and there are weekly Live Meetings where you can raise questions and interact with them. This is a technical course, aimed at programmers, system designers, and architects who want a solid understanding of the Microsoft Windows Azure platform, hence a prerequisite for this course is at least six months programming in the .NET framework and Visual Studio. Check out the full details of the event or go straight to registration.   The course outline is: Week 1 - Windows Azure Platform Week 2 - Windows Azure Storage Week 3 - Windows Azure Deep Dive and Codename "Dallas" Week 4 - SQL Azure Week 5 - Windows Azure Platform AppFabric Access Control Week 6 - Windows Azure Platform AppFabric Service Bus If you have any questions about the course and its suitability, please email [email protected].

    Read the article

< Previous Page | 171 172 173 174 175 176 177 178 179 180 181 182  | Next Page >