Search Results

Search found 31206 results on 1249 pages for 'version detection'.

Page 600/1249 | < Previous Page | 596 597 598 599 600 601 602 603 604 605 606 607  | Next Page >

  • Difference between () and [] in Perl6 [migrated]

    - by Ask and Learn
    Learning Perl 6 and had trouble to understand following Perl 6 one liner. version of Perl 6 - rakudo-star: stable 2014.04 (bottled) This works fine, array/list is sorted [njia@mb-125:~] : perl6 -e 'say [2443,5,33, 90, -9, 2, 764].sort' -9 2 5 33 90 764 2443 But this does not sort the array/list, if [].sort works why @s.sort does not? [njia@mb-125:~] : perl6 -e 'my @s = [2443,5,33, 90, -9, 2, 764]; @s.sort.say' 2443 5 33 90 -9 2 764 Change from [] to () [njia@mb-125:~] : perl6 -e 'my @s = (2443,5,33,90,-9,2,764); @s.sort.say' -9 2 5 33 90 764 2443

    Read the article

  • Updates and New Home for Sticky Plugin

    Ive created a permanent home for the latest updates for Sticky. Please visit http://johnpapa.net/sticky to stay updated on the latest versions, enhancements, and how tos for Sticky. To download the latest update go here. The latest update includes some bug fixes and added a few enhancements in the latest version: 1.0.50605. Here are the highlights: Choose your Sticky color in the Settings Numbers are formatted Better error handling for invalid data Minor graphical fixes The Settings...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • What are the advantages and disadventages of git or bzr + rsync vs rdiff-backup?

    - by Azendale
    I used to use rsync to do backups, but then I switched to rdiff-backup to incremental backups. Recently, I discovered git and bzr while working on a coding project. So, I was thinking, I could have my backup disk be a repository in either git or bzr. Then I could rsync to the repository, and commit the changes. Would there be any performance concerns with this? Any other issues that I'm not thinking of? The benefit I see in using rsync is that you can restart an interrupted transfer, while rdiff-backup reverts to the last version, and then starts again. Any reason not to do it this way? Anything I'm not thinking of?

    Read the article

  • BI Applications Mobile Demonstration

    - by Mike.Hallett(at)Oracle-BI&EPM
    Partners can now run live interactive Demos of the latest version of OBI Mobile on an iPad, and BI Applications have also been made available via OBI mobile app Demos including; Financials, HR, Marketing, Procurement & Spend, Projects and Supply chain.  You can download Demo Scripts for these: e.g. Mobile_Marketing_Analytics.pdf The mobile app is using the same dashboards and data as the BI Applications Test Drives, which partners can access here. These existing demo scripts for these BI Applications can be used with the BI mobile app.  The instructions regarding the interface will be different, but the story line is the same.  If you want the “Mobile Financial Analytics” script ask me @ [email protected] For more instructions on setting up and connecting your iPad, see: Run Live OBI Mobile HD Demos on your iPad Business doesn't stop just because you're on the go. See how Oracle BI Mobile makes consuming BI on the go simple, secure and fast.  

    Read the article

  • UFW firewall still blocking SMB despite adding rules

    - by nLinked
    I have an Ubuntu PC with ufw firewall (GUI version). I have added the preset Samba service, in and out, and even tried adding the ports manually (135-139, 445, UDP and TCP, in and out), but it still blocks samba. I am trying to access a share on another PC on my LAN. If I disable ufw, it works fine. It must be still blocking something and I can't figure it out. Any ideas? I'm on Ubuntu 11.04 beta 2.

    Read the article

  • A bacon- (and module-) saving PowerShell incident

    - by AaronBertrand
    Earlier today I made a big goof. I opened a module in Notepad, intending to use it as the basis for a new module. I was in the process of using "File > Save As" when my phone rang just at the precise instant that, for some reason, made me click on "File > Save" by mistake. After hitting Ctrl+Z 30 times to try to get the old version of the module back, I remembered that Notepad has never had more than one level of Undo. Back when I was coding ASP by hand, I was very well aware of this, but I...(read more)

    Read the article

  • Unity3D problem. Bullets fall down instead of flying like they should

    - by user2342080
    I used this tutorial as a reference. http://www.youtube.com/watch?v=3L8eaoyZ0Go My problem is that whenever I play the game, EVERYTHING works but the bullets. It just falls down instead of flying forward. This is the flash version of the game: http://v1k.me/swf/ Can some one help me out? Should I upload the project? This is my "Shoot.js": public var bulletPrefab : Transform; public var bulletSpeed : float = 20; function Update() { if(Input.GetMouseButton(0)) { if(bulletPrefab || bulletSpeed) { var bulletCreate = Instantiate(bulletPrefab, GameObject.Find("SpawnPoint").transform.position, Quaternion.identity); bulletCreate.rigidbody.AddForce(transform.forward * bulletSpeed); } } }

    Read the article

  • Installing drivers for an ATI 6950

    - by aj.esler
    Is there any way to get ubuntu to recognise an ATI 6950? I have tried installing the proprietary ATI drivers, but these do not seem to work properly. It seems to screw up the gnome desktop, and starts booting into terminal. I'm somewhat of a linux newbie, so not really sure whats going wrong. I don't want to game on ubuntu, I just want it to recognise my dual monitor setup rather than mirroring the monitors, so any driver that does this is fine by me. It seems like the open source ATI radeon drivers will not support the 6950/6970 for some time, at least until kernel 2.6.38 is released. The card is also flashed with the 6970 firmware, not sure if this is important. I am running a 64 bit version of ubuntu, I installed ubuntu server 10.10, then installed gnome-desktop.

    Read the article

  • Soft lockup after upgrade - cannot install from live CD

    - by nbm
    I dual-boot MacIntel Core 2 duo. nVidia graphics. Ran upgrade from ubuntu 13.10 to 14.04 (64 bit). On restart ran into {numbers} Bug: soft lockup - CPU#0 stuck for 22s! [swapper/0:1] Tried loading earlier kernel: same problem Tried re-installing ubuntu from a liveCD that has worked in the past: version 13.04. Same problem. Tried re-partitioning hard drive using Mac OS X disk utility and then installing ubuntu 14.04LTS from liveCD. Same problem. Not possible to verify liveCD disk (creates same "soft lockup" bug.) Tried installing from the liveCD with version 13.04 that I know works (that's how I got Ubuntu on this machine in the first place.) Same problem. I know this is not a hardware problem as OS X works just fine, I am using it right now on the same machine. I have been using various versions of Ubuntu for 2 years. Things I cannot do: Open a terminal Verify CD image Start ubuntu from CD (same soft lockup problem) This problem is similar to some other questions, none of which have been satisfactorily answered: Ubuntu 14.04 soft lockup on Vostro 3500 Cannot do fresh install of Ubuntu 13.04 while booting from DVD: "soft lockup" bug Live CD stalls when installing Ubuntu 13.10 UPDATE 6/11/14: Following some much-appreciated advice from bain (see below) I burned a 12.04LTS disk and started with kernel parameters: noapic, no1apic, acpi=off, nomodeset, elevator=deadline, and clocksource=jiffies. With all of these parameters I was able to load the 12.04LTS CD ("Try without installing"). It worked fine. However, as soon as I tried to install Ubuntu from the CD, my wired ethernet (eth0) connection would hang. There are already various askubuntu questions and bug reports about this problem, none of which had answers for me. (E.g., dhclient eth0 does nothing, none of the various reset commands does anything, manually setting IP &etc does nothing. I could reliably kill the ethernet connection by clicking "install ubuntu" every single time.) I could go ahead and install 12.04 without an internet connection, but the install would freeze after mostly completing (I tried several times.) There were some relevant error messages in the details of the install output script that, IIRC, had to do with searching for missing files and not being able to access eth0 (internet) to get them. To be honest I gave up at that point and I'm not sure I wrote those down. If I find some notes I will post them. At this point I no longer have Ubuntu on my system. I wiped the partitions and am using exclusively OS X. I am leaving this question in case it helps anyone else with similar problems. I love open source and I love Linux, and the next machine I get I will probably just build from Arch. At the moment I miss repositories and a lot of other things about Ubuntu, but the OS X terminal is 'nix, I can pretty much use all the open source apps I like, and while I am not a fan of the Apple software it gets the job done for me. Unlike Ubuntu, which can't even install. I realize this isn't necessarily a place for a soapbox speech, but when I first installed 12.04 several years ago there were already people in the community complaining that Canonical was going too "commercial". But I loved it. Several years later and all I've seen is Canonical adding more not-so-useful bells and whistles to Ubuntu while continually failing to fix basic problems on upgrades. With a dual-boot (and sometimes triple-boot) system it always took me some tweaking to get an upgrade to work, and to some extent that is okay. But at this point I feel like Canonical ought to just put a price tag on Ubuntu. All I see is more commercialism and advertising and product tie-ins, and ongoing problems do not get fixed. I am a big fan of open-source, not-for profit enterprise. I am also a big fan of for-profit enterprise, which certainly has its place and usefulness. I am not a fan of companies who pretend to be in favor of open source but really are just out to make a buck, and IMNSHO that is what Canonical has become. This is a great community and I wish you all the best, but my next install of Linux will not be Ubuntu.

    Read the article

  • wrong GNOME logo at startup after 12.04 upgrade

    - by swordfish
    I'm not too expert of Ubuntu, but after a quite painful upgrade to version 12.04 (from 11.10), which first required to reinstall wireless drivers in order to work properly, and after an update from the Update Manager that I ran soon after the upgrade, I have a very strange problem coming up: at login page, where you choose your user among the list of the available ones, I can also see the list of available IDEs (Gnome, Gnome classic, Unity, Unity 2d, etc) but the funny thing is that the classic "foot" logo normally associated to the GNOME selection has now disappeared, and instead the "gear" logo (the one normally related to Unity, is on the side of the "gnome" choice. Note that the foot logo remains associated to the Gnome Classic item, and that the "gear" logo is still associated to the various versions of Unity available. This is not serious, since I can still access my GNOME or Unity environment, but I wonder if I can fix this up.

    Read the article

  • Preventing Duplicates on Google

    - by abel
    I am currently using a rewrite rule to enable access to .php pages, without using the php extension. However to prevent old links from breaking, the pages can still be accessed via links containing the .php extension too. For eg. domain.com/page.php can now be accessed at domain.com/page All the links on the website now use domain.com/page type links within the site. However older incoming links will still link to the .php pages, meaning Google will index both pages and mark them as duplicate. I have two plans to remedy the situation. Use a php 301 redirect: When a page is accessed with the .php extension, I can redirect each page individually using a 301 redirect using php Using Canonical: Place a canonical tag on each page, pointing to the ".php" less version My Question: Are both methods equally efficacious in preventing Google from indexing my ".php" pages? Which method should be preferred, by convention or otherwise?

    Read the article

  • Yahoo arrêtera le service Yahoo Publisher Network le 30 Avril : AdSense ternit-il le marché avec sa

    Yahoo arrêtera le service Yahoo Publisher Network le 30 Avril : AdSense ternit-il le développement de ses concurrents avec sa position de leader ? [IMG]http://www.livesphere.fr/images/dvp/ypn-yahoo-publisher-network.png[/IMG] Yahoo a annoncé officiellement l'arrêt de ses services de rémunération publicitaire pour la fin avril. Le service de Yahoo n'avait pas connu un grand succès et les évolutions de la plateforme n'ont connu que peu de mises à jour et est donc reste en version Beta depuis sa mise en ligne. Ainsi, la firme recommande à leur client de se tourner vers un partenaire du nom de Chitika, pour ceux souhaitant continuer leur rémunération publicitaire. Aucune...

    Read the article

  • How do I handle "Firmware file X not found" stopping bootup?

    - by John Baber
    I recently upgraded my desktop mac to Precise and now can't get past the Ubuntu 12.04 splash page. The splash freezes with [ 19.931097] b32-phy0 ERROR: Firmware file "b43/ucode16_mimo.fw" not found [ 19.9311126] b32-phy0 ERROR: You must go to http://wireless.kernel.org/en/users/Drivers/b43#devicefirmware and download the correct firmware for this driver version. Please carefully read all instructions on this website. written on it This is when I choose 3.2.0-generic from the grub menu. If I choose to load up older kernels, I still get the same thing. How can I make my computer finish booting again? I'm able to ssh into the machine and poke around, but the actual screen freezes at the splash.

    Read the article

  • Regular wireless dropouts on new lenovo T440s in Ubuntu 14.04

    - by user290670
    Over the last 24 hours my wireless has dropped out regularly. I've tested to make sure it isn't my router (my phone and everyone in the house isn't having problems using the wifi). This is a brand-new installation of Ubuntu 14.04 and according to uname I'm running kernel 3.13.0-24-generic. Now, my laptop has an Intel 7260AC dual band wireless card and I've read that Ubuntu has been having trouble with these. I notice that at http://www.intel.com/support/wireless/wlan/sb/CS-034398.htm There are some updated drivers for my kernel version. However I have no idea how to update the kernel to use these drivers instead so that I can see if this will fix it. Can anyone help? These dropouts are really annoying. EDIT: Upgrading to 3.14.6 did not help.

    Read the article

  • Chrome 21 beta permet d'utiliser la Webcam sans plugin et intègre l'API JavaScript Gamepad

    Chrome 21 beta permet d'utiliser la Webcam sans plugin et intègre l'API Gamepad Google vient de publier la beta de la prochaine version de son navigateur Chrome. Cette mouture réduit essentiellement la dépendance du navigateur des plugins tiers comme Flash. En effet, Chrome beta apporte de nouvelles fonctionnalités qui permettront aux utilisateurs d'utiliser directement leur Webcam et Micro dans le navigateur, sans avoir recours à une extension. Cela sera possible grâce à l'utilisation de l'API getUserMedia, qui permet de récupérer simplement un flux vidéo et audio indépendamment du hardware, via la norme WebRTC. WebRTC est un framework de communications aud...

    Read the article

  • Developing Schema Compare for Oracle (Part 6): 9i Query Performance

    - by Simon Cooper
    All throughout the EAP and beta versions of Schema Compare for Oracle, our main request was support for Oracle 9i. After releasing version 1.0 with support for 10g and 11g, our next step was then to get version 1.1 of SCfO out with support for 9i. However, there were some significant problems that we had to overcome first. This post will concentrate on query execution time. When we first tested SCfO on a 9i server, after accounting for various changes to the data dictionary, we found that database registration was taking a long time. And I mean a looooooong time. The same database that on 10g or 11g would take a couple of minutes to register would be taking upwards of 30 mins on 9i. Obviously, this is not ideal, so a poke around the query execution plans was required. As an example, let's take the table population query - the one that reads ALL_TABLES and joins it with a few other dictionary views to get us back our list of tables. On 10g, this query takes 5.6 seconds. On 9i, it takes 89.47 seconds. The difference in execution plan is even more dramatic - here's the (edited) execution plan on 10g: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 108K| 939 || 1 | SORT ORDER BY | | 108K| 939 || 2 | NESTED LOOPS OUTER | | 108K| 938 ||* 3 | HASH JOIN RIGHT OUTER | | 103K| 762 || 4 | VIEW | ALL_EXTERNAL_LOCATIONS | 2058 | 3 ||* 20 | HASH JOIN RIGHT OUTER | | 73472 | 759 || 21 | VIEW | ALL_EXTERNAL_TABLES | 2097 | 3 ||* 34 | HASH JOIN RIGHT OUTER | | 39920 | 755 || 35 | VIEW | ALL_MVIEWS | 51 | 7 || 58 | NESTED LOOPS OUTER | | 39104 | 748 || 59 | VIEW | ALL_TABLES | 6704 | 668 || 89 | VIEW PUSHED PREDICATE | ALL_TAB_COMMENTS | 2025 | 5 || 106 | VIEW | ALL_PART_TABLES | 277 | 11 |------------------------------------------------------------------------------- And the same query on 9i: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 16P| 55G|| 1 | SORT ORDER BY | | 16P| 55G|| 2 | NESTED LOOPS OUTER | | 16P| 862M|| 3 | NESTED LOOPS OUTER | | 5251G| 992K|| 4 | NESTED LOOPS OUTER | | 4243M| 2578 || 5 | NESTED LOOPS OUTER | | 2669K| 1440 ||* 6 | HASH JOIN OUTER | | 398K| 302 || 7 | VIEW | ALL_TABLES | 342K| 276 || 29 | VIEW | ALL_MVIEWS | 51 | 20 ||* 50 | VIEW PUSHED PREDICATE | ALL_TAB_COMMENTS | 2043 | ||* 66 | VIEW PUSHED PREDICATE | ALL_EXTERNAL_TABLES | 1777K| ||* 80 | VIEW PUSHED PREDICATE | ALL_EXTERNAL_LOCATIONS | 1744K| ||* 96 | VIEW | ALL_PART_TABLES | 852K| |------------------------------------------------------------------------------- Have a look at the cost column. 10g's overall query cost is 939, and 9i is 55,000,000,000 (or more precisely, 55,496,472,769). It's also having to process far more data. What on earth could be causing this huge difference in query cost? After trawling through the '10g New Features' documentation, we found item 1.9.2.21. Before 10g, Oracle advised that you do not collect statistics on data dictionary objects. From 10g, it advised that you do collect statistics on the data dictionary; for our queries, Oracle therefore knows what sort of data is in the dictionary tables, and so can generate an efficient execution plan. On 9i, no statistics are present on the system tables, so Oracle has to use the Rule Based Optimizer, which turns most LEFT JOINs into nested loops. If we force 9i to use hash joins, like 10g, we get a much better plan: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 7587K| 3704 || 1 | SORT ORDER BY | | 7587K| 3704 ||* 2 | HASH JOIN OUTER | | 7587K| 822 ||* 3 | HASH JOIN OUTER | | 5262K| 616 ||* 4 | HASH JOIN OUTER | | 2980K| 465 ||* 5 | HASH JOIN OUTER | | 710K| 432 ||* 6 | HASH JOIN OUTER | | 398K| 302 || 7 | VIEW | ALL_TABLES | 342K| 276 || 29 | VIEW | ALL_MVIEWS | 51 | 20 || 50 | VIEW | ALL_PART_TABLES | 852K| 104 || 78 | VIEW | ALL_TAB_COMMENTS | 2043 | 14 || 93 | VIEW | ALL_EXTERNAL_LOCATIONS | 1744K| 31 || 106 | VIEW | ALL_EXTERNAL_TABLES | 1777K| 28 |------------------------------------------------------------------------------- That's much more like it. This drops the execution time down to 24 seconds. Not as good as 10g, but still an improvement. There are still several problems with this, however. 10g introduced a new join method - a right outer hash join (used in the first execution plan). The 9i query optimizer doesn't have this option available, so forcing a hash join means it has to hash the ALL_TABLES table, and furthermore re-hash it for every hash join in the execution plan; this could be thousands and thousands of rows. And although forcing hash joins somewhat alleviates this problem on our test systems, there's no guarantee that this will improve the execution time on customers' systems; it may even increase the time it takes (say, if all their tables are partitioned, or they've got a lot of materialized views). Ideally, we would want a solution that provides a speedup whatever the input. To try and get some ideas, we asked some oracle performance specialists to see if they had any ideas or tips. Their recommendation was to add a hidden hook into the product that allowed users to specify their own query hints, or even rewrite the queries entirely. However, we would prefer not to take that approach; as well as a lot of new infrastructure & a rewrite of the population code, it would have meant that any users of 9i would have to spend some time optimizing it to get it working on their system before they could use the product. Another approach was needed. All our population queries have a very specific pattern - a base table provides most of the information we need (ALL_TABLES for tables, or ALL_TAB_COLS for columns) and we do a left join to extra subsidiary tables that fill in gaps (for instance, ALL_PART_TABLES for partition information). All the left joins use the same set of columns to join on (typically the object owner & name), so we could re-use the hash information for each join, rather than re-hashing the same columns for every join. To allow us to do this, along with various other performance improvements that could be done for the specific query pattern we were using, we read all the tables individually and do a hash join on the client. Fortunately, this 'pure' algorithmic problem is the kind that can be very well optimized for expected real-world situations; as well as storing row data we're not using in the hash key on disk, we use very specific memory-efficient data structures to store all the information we need. This allows us to achieve a database population time that is as fast as on 10g, and even (in some situations) slightly faster, and a memory overhead of roughly 150 bytes per row of data in the result set (for schemas with 10,000 tables in that means an extra 1.4MB memory being used during population). Next: fun with the 9i dictionary views.

    Read the article

  • Can't activate "mirror display"

    - by Nemo Peeters
    I have Ubuntu 12.04 on a Dell latitude E6410, I have a problem when using the VGA output I always have dual monitor and can't switch on the "mirror display" toggle under systems setting/displays. It's rather annoying when having to do presentation with a video projector. the option "mirror display" is in light grey and I can"t tick the box. Thanks for those who might have an idea how to resolve this? update: driver is: Intel® Ironlake Mobile x86/MMX/SSE2 card something like this ? description: VGA compatible controller product: Core Processor Integrated Graphics Controller vendor: Intel Corporation physical id: 2 bus info: pci@0000:00:02.0 version: 02 width: 64 bits clock: 33MHz

    Read the article

  • Resolution Independent 2D Rendering in XNA

    - by AttackingHobo
    I am trying to figure out the best way to render a 2d game at any resolution. I am currently rendering the game at 1920x1200. I am trying scale the game to any user selected resolution without changing the way I am rendering, or game logic. What is the best way to scale a game to any arbitrary resolution? Edit: I am trying to achieve this: http://www.david-amador.com/2010/03/xna-2d-independent-resolution-rendering/ but I think the code he has is for a different version of XNA because I cannot find that method overload he uses.

    Read the article

  • Getting photos and music on/off samsung/google galaxy nexus (ice cream sandwich) phone

    - by wim
    I am having trouble to access the filesystem on my phone. It just worked in previous version of Ubuntu, but now it appears empty whether it is mounted with MTP or PTP. I have followed a few guides on building and reinstalling libmtp etc without success. This answer did not help for me, either, and gMTP just hangs when I click the connect button. I know I can use wifi e.g. airdroid to access my photos, but this is too slow and a bit clumsy for me, and downloaded photos lose their original timestamps. Has anyone had success connecting to this phone on 12.04 ??

    Read the article

  • Google intègre le support de WebM à Chrome et rechange la licence du nouveau codec vidéo issu du VP8

    Mise à jour du 07/06/10 VP8 vs H.264 : Google intègre le support du WebM à Chrome Et rechange la licence du nouveau standard vidéo issu du VP8 Les choses s'accélèrent pour le projet WebM issu du VP8, un standard vidéo que Google a décidé de rendre open-source, et du Ogg-Vorbis (lire ci-avant). Première nouvelle, Chrome intègre à présent le support du WebM. Firefox et Opera avaient déjà fait savoir qu'ils travaillaient sur le sujet. Tout comme Microsoft. La version pour développeurs du navigateur de Google (téléchargeable sur le dev channel) permettra donc à tout un chacun de se faire une opinion personnelle sur les qua...

    Read the article

  • Content, MetaData and Taxonomy 2 Overview of the Data Layer

    This article is cross-posted from my personal blog. In DotNetNuke version 5.3, we introduced the concept of a centralized Content store, together with the ability to apply Taxonomies (categories) to the content. We have extended this in DNN 5.4 by completing the MetaData API as well as adding Folksonomy (user tags). In this series of blogs I will explain how developers can take advantage of these new features in their own extensions. In the first blog in this series I covered the Taxonomy Manager...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • What does the kernel boot parameter "set acpi_osi=Linux" do?

    - by AIB
    I am facing an issue with brightness control in my laptop which has Intel HD graphics. So when I searched for a fix I found that a common suggestion is to edit the GRUB file line, GRUB_CMDLINE_LINUX_DEFAULT="quiet splash" to GRUB_CMDLINE_LINUX_DEFAULT="quiet splash acpi_osi=Linux acpi_backlight=vendor" So could anyone tell me what the kernel boot option acpi_osi=Linux do? My OS is Ubuntu 10.10 64-bit version update The boot parameter acpi_osi is used as follows, acpi_osi= [HW,ACPI] Modify list of supported OS interface strings acpi_osi="string1" # add string1 -- only one string acpi_osi="!string2" # remove built-in string2 acpi_osi= # disable all strings But still I don't find a reason why this should fix any laptop issue!!

    Read the article

  • International TLD's vs. duplicate content

    - by Litso
    Hey all, I currently work at a pretty big website that has visitors from around the globe. My job is to help out on the SEO, and one thing we've been discussing lately is the use of international TLD's. The ones we use range between: (partly) translated websites like .es and .de that serve most of the content in the country's language non-translated (english) websites for non-english languages (due to a lack of translations) like .ro and .cz english websites for english speaking countries with localized TLD's (.co.nz, .co.uk) On one hand I really have the feeling this is causing a lot of duplicate content, especially for the last two categories of TLD's. On the other hand though it seems a lot like country-specific TLD's tend to score a lot better in that country's Google. Would it be advisable to keep on using these domains, or should we canonicalize them all to the .com version?

    Read the article

  • "From the Coal Face" - 1 - What ILDASM can reveal!

    - by TATWORTH
    In a place far, far away, there was a project where the Architect decided on using embedded TSQL in a Dot Net application, rather than use stored procedures. I located ILDASM.EXE (my Framework 3.5 version lives at C:\Program Files\Microsoft SDKs\Windows\v7.0A\bin on my home development PC) and created a shortcut to it in the SendTo folder. Now I could set about doing a simple demonstration to the Architect by taking one of the Dot Net EXE's with the embedded TSQL and sending it to ILDASM.  Since I had written most of the embedded TSQL, it was a matter of seconds before I located the embedded TSQL within the Exe. The TSQL that was supposed to be safely hidden within the EXE was easily located and and copied. (It should also be noted that we could have encrypted the stored procedures on loading them to the database.)

    Read the article

  • Ubuntu pense à remplacer Firefox par Chromium dans sa distribution pour Netbooks : voteriez-vous pou

    Ubuntu teste Chromium pour remplacer Firefox Dans sa future distribution pour Netbooks : voteriez-vous pour Google ou pour Mozilla ? Ubuntu envisage d'abandonner Firefox pour sa version Netbook et de le remplacer par Chromium, l'implémentation libre du navigateur de Google. La décision n'est pas encore définitive. Mais elle indique clairement que l'équipe de développement de la distribution Linux la plus connue du grand public est en train de se lasser des « lourdeurs » du navigateur de Mozilla. C'est en tout cas l'idée exprimée lors depuis quelques temps déjà : « Firefox est un navigateur web assez lourd qui souffre de performances insuffisamment optimisé...

    Read the article

< Previous Page | 596 597 598 599 600 601 602 603 604 605 606 607  | Next Page >