Search Results

Search found 31026 results on 1242 pages for 'google cloud platform'.

Page 323/1242 | < Previous Page | 319 320 321 322 323 324 325 326 327 328 329 330  | Next Page >

  • Bridging Virtual Networking into Real LAN on a OpenNebula Cluster

    - by user101012
    I'm running Open Nebula with 1 Cluster Controller and 3 Nodes. I registered the nodes at the front-end controller and I can start an Ubuntu virtual machine on one of the nodes. However from my network I cannot ping the virtual machine. I am not quite sure if I have set up the virtual machine correctly. The Nodes all have a br0 interfaces which is bridged with eth0. The IP Address is in the 192.168.1.x range. The Template file I used for the vmnet is: NAME = "VM LAN" TYPE = RANGED BRIDGE = br0 # Replace br0 with the bridge interface from the cluster nodes NETWORK_ADDRESS = 192.168.1.128 # Replace with corresponding IP address NETWORK_SIZE = 126 NETMASK = 255.255.255.0 GATEWAY = 192.168.1.1 NS = 192.168.1.1 However, I cannot reach any of the virtual machines even though sunstone says that the virtual machine is running and onevm list also states that the vm is running. It might be helpful to know that we are using KVM as a hypervisor and I am not quite sure if the virbr0 interface which was automatically created when installing KVM might be a problem.

    Read the article

  • Can't download web photo albums to Picasa

    - by Arcadie
    Someone has shared a Picasa web album (Limited, anyone with the link), but I can't download it to Picasa. The following alert appears: Firefox doesn't know how to open this address, because the protocol (picasa) isn't associated with any program. I have Picasa 3.0.0 installed on Ubuntu 11.04, I remember it saying something about registering the picasa protocol with Firefox during the installation. I have Firefox 6.0.2, and these settings are present in about:config network.protocol-handler.app.picasa;/usr/bin/picasa network.protocol-handler.expose.picasa;true network.protocol-handler.external.picasa;true Picasa is located here: $ which picasa /usr/bin/picasa Is there something I can do to make this work? PS: I hope this is not off-topic here, and I can't find the "picasa" tag. Could someone please add it, if appropriate?

    Read the article

  • When I try to click and launch some of the links set to open in new window, it is being treated as a pop-up window [migrated]

    - by Test Developer
    For the past few days, we are facing issue with the chrome browser behavior. This is related to opening links set to open in new tab/window. The details are as follow: I have a collection of links and each link points to different resource to be opened in a new tab/window. The code is as follow: <a class="cssClass" rid="1114931" href="http://www.domain.com/resources/abc.html" title="Link1" tabindex="4">Link 1</a> And there are few checks/filters over accessing the resources which have been implemented as onClick handler over the links. In case any of the validations fails, the onClick handler returns false and the default behavior of the link does not happens i.e. links does not get open. One of such (last) checks includes AJAX call in sync mode. The code is as follow: var link_clickHandler = function(evt/* Event */) { var objTarget = jQuery(evt.target); if(check1) { return false; } else if(check2) { return false; } else if(check3) { var blnRetVal = false; jQuery.ajax( { "async" : false, "type" : "GET", "contentType" : "application/json; charset=utf-8", "url" : "index.php", "data" : 'resourceid=' + intResourceId, "dataType" : "json", "forceData" : true, "success" : function(data) { if(check1) { blnRetVal = true; } } "error" : function(error) { } } ); return blnRetVal; } }; jQuery("a.cssClass").live("click", link_clickHandler); ISSUE: The issue is that Chrome is behaving very weirdly manner. In case all of the checks are passed and onClick handler returns true, sometimes the resource get opened in a new tab/window and sometimes it get opened as a pop-up (which should never). Tried to capture any pattern but could not succeed. Any solution or even helping in understanding behavior would be really appreciated.

    Read the article

  • The Beginner’s Guide To Tabbed Browsing

    - by Chris Hoffman
    Tabs allow you to open multiple web pages in a single browser window without cluttering your desktop. Mastering tabbed browsing can speed up your browsing experience and make multiple web pages easier to manage. Tabbed browsing was once the domain of geeks using alternative browsers, but every popular browser now supports tabbed browsing – even mobile browsers on smartphones and tablets. This article is intended for beginners. If you know someone that doesn’t fully understand tabbed browsing and how awesome it is, feel free to send it to them! How Hackers Can Disguise Malicious Programs With Fake File Extensions Can Dust Actually Damage My Computer? What To Do If You Get a Virus on Your Computer

    Read the article

  • Automatically Login & Startup A Windows Program On Amazon's EC2 Service

    - by darkAsPitch
    How can I automatically start a program on Amazon's EC2 Windows 2008 web servers? For example, if I wanted to test the "Digg effect" on a web page of mine, how could I open 100 windows 2008 servers at once, each loading one (or two?) instances of the firefox web browser? I have placed a sample batch file in the windows startup folder that echos the time it was called, but it is only started when I actually login remotely via the remote desktop protocol. I don't want to have to login to 100 servers to get my software to run :P What can I do? I am using this Windows 2008 Datacenter, Amazon-supplied AMI specifically: ami-a2698bcb

    Read the article

  • Two hosting providers running simultaneously... possible / not possible? good practice / unnecessary?

    - by user29600
    For the sake of their reputation, I won't mention the names. But I'll just use: Business I worked for previously - ABC Web Dev Hosting company they used - XYZ Hosting I recently found out that XYZ Hosting had some sort of incident where they ended up losing a lot of their client's data - including ABC Web Dev's. ABC Web Dev was able to recover some of their customer's websites, after pulling them from their local development computers and putting them up on another hosting provider. They ended up losing a lot of clients because of it and their reputation ruined. I'm starting my own web dev company and I don't want to run into this same issue. I'm planning on using Rackspace but, although they are a great company, according to wikipedia they still have had downtime in their past. I thought it might be a good idea to try to run two providers at once, to ensure that if anything happened in one the websites would still be live because of the other. I know the websites would have to be pulling from one server at all times, but if there's a way to redirect requests to the second server if the first one is down that would solve my issue. As a note, we will have a staging environment setup locally which will allow for quick recovery if a provider did have any issues, however I'd like to avoid any downtime at all if possible. So my questions are: Has anyone tried running two providers simultaneously? Would this be considered good practice or am I going too far? Is there really any way to run two simultaneously where one server acts as a backup?

    Read the article

  • Can't find standalone Chrome Gmail client that I know exists

    - by Carson
    I'm on Windows. A couple years ago when I switched from Outlook to Gmail (Google Apps), Google provided this awesome little standalone gmail client that was just a single-purpose Chrome install. It launched like a normal application, stayed updated when I updated Chrome. It was Chrome in a separate application that launched only gmail, stayed logged in really well, and "felt" like a gmail mail client, with the gmail interface. It had it's own little red envelope icon, it was a windows app. (I remember there was no Mac equivalent.) I found it while looking through the "this is how you get your company to switch to gmail" documentation that Google provided. I just repaved my box and now I'm looking for this thing again, and I had no idea it would be impossible to find. I've spent literally 2 hours looking, searching, googling, etc. I'm losing my mind. Anyone know how I can get my hands on this? I used it all day every day for 2 years, so I know it exists :), but I can not find it. Any assistance would be gratefully received.

    Read the article

  • The number of soft 404 errors is increasing because of redirects to the home page

    - by Stevie G
    I have an increase in soft 404 errors. Using Apache in my .htaccess file I have: Redirect 301 /test.html? /page/pop/test Redirect 301 /about.html? /about I have also tried: Redirect 301 http://www.example.co.za/test.html? http://www.example.co.za/services/test however whenever I go to: http://www.example.co.za/test.html http://www.example.co.za/about.html it just redirects to the home page I also have: RewriteRule ^.*$ index.php [NC,L] in .htaccess

    Read the article

  • Build tools for php, html, css, js web app development

    - by cs_brandt
    What are some recommendations for a build tool that would allow me to upload changes to a web server or a repository and minify the js and css automatically, and possibly even run Closure compiler on the JavaScript? Im not worried about doing anything with the php code other than update with most recent changes although in the future would like to have phpdoc updated automatically. Just wondering if there is some way to do all this other than an amalgam of scripts that run or have to be invoked every time. Thanks.

    Read the article

  • Google présente les trois stratégies les plus importantes pour le référencement en 2011 : rapidité, contrôle interne et réseaux sociaux

    Google présente les trois stratégies les plus importantes pour le référencement en 2011 Rapidité, contrôle interne et Webmarketing sur les réseaux sociaux Sur sa chaine YouTube, le centre Google Webmaster Help répond régulièrement aux questions les plus pertinentes venant de professionnels des divers métiers du web. Cette semaine, Matt Cutts, responsable de l'équipe anti-spam de Google, répond à une question particulièrement intéressante : « Si vous étiez un expert en optimisation pour les moteurs de recherche dans une grande entreprise, quelles sont les trois choses que vous incluriez dans votre stratégie pour 2011 ? » Pour Matt Cutts, la première chose qu'i...

    Read the article

  • Des ingénieurs sécurité Google disent «Fuck you» à la NSA, après les révélations du programme d'écoute Muscular

    Des ingénieurs sécurité Google disent « Fuck you » à la NSA, après les révélations du programme d'écoute Muscular Les révélations relatives aux programmes d'écoute de la NSA semblent avoir eu un impact sur ces poids lourds de la technologie qui étaient eux aussi mis en cause. Chez Google par exemple, Eric Schmidt et David Drummond, respectivement dirigeant de la firme et avocat en chef de Google, s'étaient dits « scandalisés ». Mais deux ingénieurs de l'entreprise se sont illustrés par des...

    Read the article

  • Quelle firme représente le plus grand danger pour Google : Facebook ou Microsoft ? Eric Schmidt penche pour le second choix

    Quelle firme représente le plus grand danger pour Google : Facebook ou Microsoft ? Eric Schmidt penche pour le second choix Depuis quelques mois, les citations et rapports avançant que Facebook est le plus grand rival de Google ne cessent de se succéder. Mais est-ce vraiment le cas ? Pas au goût d'Eric Schmidt en tous cas, le CEO sortant de la firme de Mountain View. Selon lui, l'entreprise faisant le plus de concurrence à Google est Microsoft. Il précise que Redmond possède plus de cash, d'ingénieurs et de clients, ce qui pousse Google à "sentir la concurrence de Microsoft chaque jour". Alors que, parallèlement à cela, Facebook a clairement expliqué qu'il ne s'attaquerait pas ...

    Read the article

  • Variable size encrypted container

    - by Cray
    Is there an application similar to TrueCrypt, but the one that can make variable size containers opposed to fixed-size or only-growing-to-certain-amount containers which can be made by TrueCrypt? I want this container to be able to be mounted to a drive/folder, and the size of the outer container not be much different from the total size of all the files that I put into the mounted folder, while still providing strong encryption. If to put it in other words, I want a program like truecrypt, which not only automatically grows the container if I put in new files, but also decreases it's size if some files are deleted. I know there are some issues of course, and it would not work 100% as truecrypt, because it basically works on the sector level of the disk, giving all the filesystem-control to the OS, and so when I remove a file, it might as well be left there, or there might be some fragmentation issues that would stop just truncating the volume from working, but perhaps a program can be built in some other way? Instead of providing sector-level interface, it would provide filesystem-level interface? A filesystem inside a file which would support shrinking when files are deleted?

    Read the article

  • Create an AWS HVM Linux AMI from an Existing Paravirtual Linux AMI

    - by javacavaj
    Is it possible to create a hardware virtual machine (HVM) AMI from an existing paravirtual (PV) AMI. My initially thought was to start a new PV instance and use the ec2-create-image command to create a new image while specifying HVM as the virutalization type. However, ec2-create-image does not have a command line parameter to specify the virtualization type. Is there another way to go about doing this?

    Read the article

  • Memcached server: Is it a good practice to point two server urls to the same server?

    - by Niro
    I have a system where there are connections to a memcache server from several different files and servers. I would like to stay with one server but keep the option of increasing the number of memcache servers (for periods of of high traffic). My idea is to tell memcache there are two servers, while the two urls will point (by DNS) to a single server. In the future if I want I can add a server and change DNS without changing the code in many places. Is this a good practice? Is there a performance cost to the fact that there are two server connections but they both point to the same server? Any other idea how to achive instant expeandability of memcache capacity without need to change the code and deploy ?

    Read the article

  • higher cpu usage in ubuntu 12.04 than windows 7

    - by Medya
    Hi I have a Intel Core i5 with 6GB of Ram using ubuntu 12.04 64bit. I noticed that whenever I run chrome which is the faster browser for me in linux) when I watch youtube, the CPU usage for Chrome is at least 13% and sometimes even 30% . but in Windows 7 same thing (youtube on chrome) rarely uses more than 6% of my CPU usage. I also notice my laptop is so hot in ubuntu 12.04 and the fan is working all the time, while in windows the laptop is so silent and the fan doesnt make much noise all the time and not as warm as in linux. is it like that for every one or is it just me?

    Read the article

  • File sharing for small, distributed, non-technical, non-profit organization?

    - by mnmldave
    Problem: I've started volunteering for a small non-profit with fewer than five non-technical Windows users who need to share 20-30GB of files (Office documents, images, PDFs, etc.) amongst themselves online. Background: The users are accustomed to a Windows network share on a machine that backed up their data locally. An on-site "disaster" has forced them to work from their homes for awhile and to re-evaluate their file sharing needs (office was located in an old building with obvious electrical issues, etc.). Access to time from volunteers with IT experience seems to be difficult. Demonstrably minimizing energy consumption is a nice-to-have. I'm currently considering Jungle Disk (a Desktop account shared amongst the handful of employees since their TOS and my inquiries to their helpdesk seem to indicate this is permissible). It appears easy-to-use, inexpensive, secure, has backup functionality, and can scale to accomodate more data when needed. I've not used it myself though (have only used Dropbox for personal use) and systems isn't my area of expertise, so am worried I might be jumping on a bandwagon. That said, any suggestions, thoughts or similar experiences would be really appreciated.

    Read the article

  • NGiNX + PHP5-fpm + CDN Tools (plugin)

    - by chris hough
    I am trying to activate the CDN tools plugin and I keep getting the following error: Fatal error: Allowed memory size of 67108864 bytes exhausted (tried to allocate 30720 bytes) in /srv/www/www.triathleteskitchen.com/wp-content/plugins/cdn-tools/cdn_classes/cloudfiles/cloudfiles_http.php on line 252 After extensive research on this issue in which I updated both of the following settings: max_execution_time = 300 memory_limit = 128M and verified the settings are active by setting up a dump phpinfo() page. Still no luck /cry I am curious if any php geeks better than me have any ideas or can point me in the right direction. Happy Holidays to you and your families :)

    Read the article

  • EC2 Configuration

    - by user123683
    I am trying to create a server structure for my EC2 account. The design I have chosen consists of 2 instances running in different availability zones, elastic load balancer, an auto-scaling group with cloudwatch monitoring configured and a security group defining rules for access to the instances. This setup is to support an online web application written in PHP. I am trying to decide what is a better policy: Store MySQL DB on a separate Instance Store MySQL DB on an attached EBS volume (from what i know auto-scaling will not replicate the attached EBS volume but will generate new instances from a chosen AMI - is this view correct?) Regards the AMI I plan to use a basic Amazon linux 64 bit AMI, and install bastille (maybe OSSEC) but I am looking to also use an encrypted file system. Are there any issues using an encrypted file system and communication between the DB and webapp i neeed to be aware of? Are there any comms issues using the encrypted filesystem on the instance housing the webapp I was going to launch a second instance or attach a second volume in the second availability zone to act as a standby for the database - I'm just looking for some suggestions about how to get the two DB's to talk - will this be a big task Regards updates for security is it best to create a recent snapshot and just relaunch and allow Amazon to install updates on launch or is the yum update mechanism a suitable alternative - is it better practice to relaunch instead of updates being installed which force a restart. I plan to create two AMI snapshots one for the app server and one for the DB each with the same security measures in place - is this a reasonable - I just figure it is a better policy than having additional applications that are unnecessary included in a AMI that I intend on using. My plan for backup is to create periodic snapshots of the webapp and DB instances (if I use an additional EBS volume instead of separate instances my understanding is that the EBS volume will persist in S3 storage in the event of an unexpected termination and I can create snapshots of the volume backup purposes). Thanks in advance for suggestions and advice. I am new to EC2 and I may have described unnecessary overkill but I want to try implement what can be considered a best practice solution so all advice is appreciated.

    Read the article

  • What is the maximum number of characters in the utm_content param in GA?

    - by David Parks
    For example, we want to differentiate people who followed our daily product email blast. I could use the product ID for utm_content, but it would be easier to read to use the SEO friendly URL path, such as: http://www.oursite.com/products/really-great-new-product https://www.frugg.com/? utm_source=a&utm_medium=b& utm_term=c& utm_content=Can-I-use-a-really-long-content-tag-like-this-one-or-is-this-going-to-break-something& utm_campaign=d

    Read the article

  • Google s'achemine vers la recherche sans recherche ou comment délivrer des résultats sans que l'utilisateur n'ait à le demander

    Google s'achemine vers la recherche sans recherche Ou comment délivrer des résultats sans que l'utilisateur n'ait à le demander Lors de la conférence « LeWeb10 », qui se tient actuellement à Paris (à Saint-Denis exactement), Marissa Mayer, ex-Vice Présidente du Search Product (du moteur de recherche donc) de Google - et à présent en charge de la branche Géolocalisation et des Services Localisés - est revenue sur une des perspectives d'avenir de la société. Lors d'un entretien sur la grande scène de la manifestion, elle a ainsi laissé entendre qu'un des prochains produits importants de Google serait la « découverte contextuelle » (contextual discovery) Il ne s...

    Read the article

  • How Do I Export Pages from Browser with Embedded Hyperlinks?

    - by Volomike
    Made a sad discovery today. I have Ubuntu 10.04 LTS. My client is in the ad business and she had a marketing competition task for me. She wanted me to visit websites of the competitors, and export the home pages as PDF. However, she wanted me to do so with embedded hyperlinks. As it turns out, Firefox (and even the latest Chrome) on Ubuntu 10.04 LTS do not embed hyperlinks in PDF web page exports. Sure, there are several Chrome and FF plugins that let you export as PDF, but what these do is connect to the URL remotely, generate the PDF remotely, and then force a download in your browser to download it from a remote location. That's not good for me, though, because some of these competitor pages require an initial login. That means that all I get back on the PDF printing from these FF or Chrome plugins is a login page. Is there a way to get around this problem, to fix the broken PDF printer on Ubuntu 10.04?

    Read the article

  • How to install the latest version of google earth on ubuntu 12.10 64-bit?

    - by user114769
    Chris here with a huge problem with the Google earth latest version. Gosh.. whenever I try to lunch the app this comes out. I'v Tried this web site nothing worked: (THANK YOU SO MUCH FOR EVEN READING THIS. MAY YOU GUYS HAVE A NICE DAY. christopher@christopher-E4300:~$ google-earth Google Earth has caught signal 11. We apologize for the inconvenience, but Google Earth has crashed. This is a bug in the program, and should never happen under normal circumstances. A bug report and debugging data have been written to this text file: /home/christopher/.googleearth/crashlogs/crashlog-50cbd67e.txt Please include this file if you submit a bug report to Google. https://help.ubuntu.com/community/GoogleEarth#Installing_the_.deb_file_downloaded_from_the_Google_Earth_Website Here is the content of /home/christopher/.googleearth/crashlogs/crashlog-50cbd67e.txt Major Version 7 Minor Version 0 Build Number 0001 Build Date Oct 29 2012 Build Time 19:13:39 OS Type 3 OS Major Version 3 OS Minor Version 5 OS Build Version 0 OS Patch Version 0 Crash Signal 11 Crash Time 1355535998 Up Time 0.789556 Stacktrace from glibc: ./libgoogleearth_free.so(+0x1e9cfb)[0xf757dcfb] ./libgoogleearth_free.so(+0x1e9f43)[0xf757df43] [0xf7726400]

    Read the article

  • Best option for storage clustering

    - by sam
    I'm working on an application that requires a large amount of storage space and I want to handle storage 'in-house' (Much cheaper than, say, S3) so we will have multiple servers (Initially 4) with large amounts of storage (6TB each). The storage will need to be very flexible and configurable, each piece of data should be replicated on at least 2 servers and must be easily readable/writable from ether an API of a UNIX device/file/folder like a normal drive, I don't mind which. We must also be able to easily offload content to our HTTP CDN (Edgecast), it doesn't need to have built in HTTP support but if it doesn't I'm going to have to write something to get the files onto HTTP so they can be pulled by the CDN. I've looked at a lot of solutions including Eucalyptus Walrus OpenStack Object Storage MogileFS and some others which I can't remember All the servers will be running RHEL 6, they have 4x1.5TB drives which will be RAID1'd into a single partition. All the servers have 1GB/s connections between them and 100MB/s connections to the internet with unlimited bandwidth. They have 2x2.66ghz processors. I understand there isn't a single, perfect answer but it would be nice to get some pointers.

    Read the article

  • Chrome Rewrite of Host: in HTML GET

    - by user912679
    At some point in the past I had a plugin for firefox that rewrites the HTML headers being sent by your browser, specifically the "Host:" line in the HTML GET request. I can't find this plugin online. Does anyone know a plugin/way to do this? I am looking for one for Chrome but any would work. The specific reason for this is I am trying to work on a wordpress website which I just did a DNS change on. Until that DNS change goes into effect I can use the IP but since its a shared host the Host line isn't set right.

    Read the article

< Previous Page | 319 320 321 322 323 324 325 326 327 328 329 330  | Next Page >