Search Results

Search found 11025 results on 441 pages for 'f4r 20'.

Page 152/441 | < Previous Page | 148 149 150 151 152 153 154 155 156 157 158 159  | Next Page >

  • How to deal with characters picking up and dropping objects in a 2D game

    - by pm_2
    I'm quite new to game development, so would like to get a consensus on methods of doing this. My game features a 2D character that is able to pick up and drop objects, for example, a stick. My question is: is it advisable / possible to manipulate the image of the character and image of the stick to make it look like the character is now carrying a stick; or is it best to have a separate sprite sheet for the character with the stick and the character without? EDIT: To be clear - I have a lot of characters, with a few items (4 separate items and over 20 characters)

    Read the article

  • SQL Saturday #310 - Dublin, Ireland

    SQL Saturday is coming to Dublin on September 20, 2014. Come for a free day of SQL Server training and networking. This year's conference features a mix of levels, topics, and speakers like Buck Woody (Big Data), Jen Stirrup (PowerBI), Denny Cherry (Storage), Red Gate's Tom Austin (Continuous integration), and more. Register while space is available. Need to compare and sync database schemas?Let SQL Compare do the hard work. ”With the productivity I'll get out of this tool, it's like buying time.” Robert Sondles. Download a free trial.

    Read the article

  • I can not download anything

    - by Jason Machen
    I am very new to ubuntu but decided to wipe my windows 7 and install it. I can not download anything from the software center. This is the error message I get. I can use the web in all other ways including this site. What can I do? Thanks, Jason W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/main/source/Sources 404 Not Found [IP: 91.189.91.13 80] W:Failed to fetch http://security.ubuntu.com/ubuntu/dists/raring-security/restricted Plus about 20 other lines.

    Read the article

  • How to divide hex grid evenly among n players?

    - by manabreak
    I'm making a simple hex-based game, and I want the map to be divided evenly among the players. The map is created randomly, and I want the players to have about equal amount of cells, with relatively small areas. For example, if there's four players and 80 cells in the map, each of the players would have about 20 cells (it doesn't have to be spot-on accurate). Also, each player should have no more than four adjacent cells. That is to say, when the map is generated, the biggest "chunks" cannot be more than four cells each. I know this is not always possible for two or three players (as this resembles the "coloring the map" problem), and I'm OK with doing other solutions for those (like creating maps that solve the problem instead). But, for four to eight players, how could I approach this problem? As always, any and all help is appreciated. :)

    Read the article

  • Will the Global Demand for Water Outstrip the Supply by 2030?

    - by Evelyn Neumayr
    A recent study conducted by the Economist Intelligence Unit and sponsored by Oracle Utilities, titled “Water for All?”,considers the preparedness of utilities to supply water to the current global population of over 7 billion people, with a further 1 billion expected by 2030. It compares strategies used by utilities in 10 major countries to address this challenge. This study’s findings show that wide-ranging water management efforts and large-scale investments must be made if utilities are to meet near-certain water stress—demand outstripping supply—by 2030. The report is based on an online survey of 244 executives of water utilities in these countries, supplemented by in-depth interviews with 20 water utility executives and independent experts. The research concludes that utilities worldwide expect to meet future demand, despite increased supply pressure on supplies, due to improvements in water productivity that the wide range of measures utilities and governments will take to ensure that water is used more efficiently. Read more about this here.

    Read the article

  • How can I make a case for "dependency management"?

    - by C. Ross
    I'm currently trying to make a case for adopting dependency management for builds (ala Maven, Ivy, NuGet) and creating an internal repository for shared modules, of which we have over a dozen enterprise wide. What are the primary selling points of this build technique? The ones I have so far: Eases the process of distributing and importing shared modules, especially version upgrades. Requires the dependencies of shared modules to be precisely documented. Removes shared modules from source control, speeding and simplifying checkouts/check ins (when you have applications with 20+ libraries this is a real factor). Allows more control or awareness of what third party libs are used in your organization. Are there any selling points that I'm missing? Are there any studies or articles giving improvement metrics?

    Read the article

  • How can I motivate the community to use boinc?

    - by user49523
    boinc is a really nice application to use on background of a computer that can help science projects and humanity as a all. I would like somehow to encourage people use boinc with their preferences on boinc like projects choice, cpu from 10 to 100 ,.. with their computer that will help science to move more quickly and help humanity There are 20 million users of ubuntu if most of use boinc , this will increase the speed of science projets .. and if other people from other operating systems use as also like mac,windows,.. and other open source like debian,.. this would even speed more science and now anyone can win also a prize , joining a project management like http://boincstats.com/ and there is more 2 How can i get ubuntu users and ubuntu community to know and use boinc application?

    Read the article

  • How to organize the nautilus bookmarks in the "Places" panel applet?

    - by piedro
    their seems to be no configuration dialog for the "Places" menue in the panel. Yes, I know how to add, remove or change the order of the bookmarks in nautilus. But I want to use folders and subfolders for the bookmarks. With more than 20 entries the nautilus bookmarks and the places menue become inconvenient. any editor for this? any configuration file that does the job? any other tool than the standard places menue? any extension for nautilus to extend the bookmark organization? thx for reading, p.

    Read the article

  • Un nouveau virus d'une complexité hors-norme mis à jour, ses auteurs sont tout aussi mystérieux que ceux de son cousin Stuxnet

    Un nouveau virus d'une complexité hors-norme mis à jour Flame serait un projet parallèle à Stuxnet dont les auteurs sont tout aussi mystérieux Flame, Skywiper ou Flamer. Tels sont les trois noms que les différents cabinets d'experts en sécurité informatique ont donné au « nouveau » virus qu'ils ont récemment découvert. Un virus d'une taille (20 Mo) et d'une complexité hors norme. Plusieurs de ces caractéristiques font fortement penser au désormais célèbre Stuxnet et laissent à penser que ses commanditaires sont identiques. C'est en tout cas ce que rapporte le Washington Post. « Il est très probable que deux équipes aient travaillé sur le même programme mais avec deux ap...

    Read the article

  • What is the term for a really BIG source code commit?

    - by Ida
    Sometimes when we check the commit history of a software, we may see that there are a few commits that are really BIG - they may change 10 or 20 files with hundreds of changed source code lines (delta). I remember that there is a commonly used term for such BIG commit but I can't recall exactly what that term is. Can anyone help me? What is the term that programmers usually use to refer to such BIG and giant commit? BTW, is committing a lot of changes all together a good practice? UPDATE: thank you guys for the inspiring discussion! But I think "code bomb" is the term that I'm looking for.

    Read the article

  • What is the reason for high power consumption in 12.04?

    - by tom
    I haven't seen this exact question posted or any related answers, so I'm re-posting. Here is the problem: After upgrading to Ubuntu 12.04 Precise Pangolin, my t420s laptop idles above 20 watts (right now with only Chrome running, I'm using 25.4 W) I had a similar problem with Ubuntu 11.10, but after much tweaking the power consumption came down < 10 W on idle. The primary culprit to the 11.10 problem was supposedly fixed by default in 12.04. So my question is, what is happening now? Computer: Lenovo Thinkpad t420s, with Intel i5-2520M @2.5 Ghz - 2x 4gb ram - disk 0 HITACHI 320 Gb - disk 1 SATA SSD 128 Gb

    Read the article

  • GDD-BR 2010 [1B] What's New in Google App Engine and GAE for Business

    GDD-BR 2010 [1B] What's New in Google App Engine and GAE for Business Speaker: Patrick Chanezon Track: Cloud Computing Time: B[11:15 - 12:00] Room: 1 Level: 151 Learn what's new with Java on App Engine. We'll take a whirlwind tour through the changes since last year, walk through a code sample for task queues and the new blobstore service, and demonstrate techniques for improving your application's performance. We'll top it off with a glimpse into some new features that we've planned for the year ahead. This session will include an overview of Google App Engine for Business. From: GoogleDevelopers Views: 0 0 ratings Time: 49:20 More in Science & Technology

    Read the article

  • HFS+/How to convert an XFS file system to HFS+ [on hold]

    - by user219350
    I have repeatedly convinced of the reliability of the XFS file system , and I was more than satisfied . I was happy with everything in Ubuntu 14.04 ( great software) , but there is a little "but ! " Basically, I work on OSX-Mavericks 10.9.3, which sees very Windows 8.1 and works wonders with NTFS, but Ubuntu does not see! Briefly describe the equipment: ASRock B75 Pro3-M i5 3330 GeForce GTX 650 Ti SATA 500GB running OS X Mavericks + Clover - a boot disk Toshiba 2TB running Windows 8.1 (x64) and Ubuntu 14.04 (amd64) If you boot from the Toshiba (where there is Ubuntu and boot Windows + GRAB) after restart boot from Clover, it is impossible. Tried a lot of options - as Clover installation and boot priority, and various settings Grab, but have not found an acceptable option and reinstall again Clover - no desire ( Mavericks 20 seconds reboots - excellent !) So please help on the file system - how to convert to XFS HFS + magazine . Mavericks to saw it all synced on Mac . Thank you for the sensible answer and help! Originally in Russian.

    Read the article

  • Why are downloads from Canonical Partners repository so slow?

    - by Sabacon
    If I need Sun Java, Adobe Flash Plugin or anything else that comes from Canonical Partners the package downloads are painfully slow even small sized packages like the Flash plugin, to speed things up I have to go here: http://archive.canonical.com/ubuntu/pool/partner/ to find what I want, download the packages with a download manager (which is usually about 20 times faster than the package manager) and then place them in my /var/cache/apt/archives folder I run the package manager afterwards, as long as the right versions of the packages I ask to install are detected in the /var/cache/apt/archives folder they will be installed immediately. I would like to stop doing this, so I am wondering if anyone else has this problem, what could be the cause and if there is a fix. I am located in the Western Caribbean region. I think it would be helpful to note that all other packages coming from the repository I have selected with synaptic download at acceptable speeds.

    Read the article

  • How to read data from a large number of files in a folder? [closed]

    - by Gary Dhillon
    I seem to be having some trouble figuring out a solution for a problem. See the thing is, my code is supposed to read a lot of data from a bunch of files. I've been thinking of two different approaches: 1) the first one seems simpler, I ask the user if they would like to examine the next file or just quit out of the program.( I believe this is simpler and would take less time to run through.) 2)It reads through all the files and outputs the results for each of them, and then a shared result for all of them.( I think this would be better for what I've been asked to do and it saves the user some hassle.) If anyone can tell me how to code either of these in C++, I would be very grateful. Here is a sample of the file: 0 -- 19 weight 0 -- 20 weight I use this to determine density and possibly ignore the weights which is a number.

    Read the article

  • MacMini (running Ubuntu 14.04) loses wlan connection when uploading larger files (several 100 mb) via ownCloud

    - by ManekenT
    I installed Ubuntu 14.04 on an old MacMini with the intention of running it as a homeserver. Additionally I installed ownCloud and tried to sync some files both from a laptop running elementaryOS and a desktop running windows 7. Syncing smaller files workes like a charm (4000 files at <10mb each) but when it comes to bigger files (1 GB ubuntu iso e.g.) the upload failes after 20-100mb. I can't ping the server anymore and the server can't ping me. It still shows up in our router as connected though. Disconnecting and reconnecting the wlan connection fixes the issue until the next attempt at syncing. Edit: I also had to install the wlan driver with this manual: https://help.ubuntu.com/community/MacBookPro8-2#Wireless

    Read the article

  • In Google Analytics, how can I determine the value of a page if no goals or revenue have been determined?

    - by Brandon Durham
    I have 4 years of data in Analytics with over 20 million pageviews for the entire site. No goals have ever been set up, and while the site is an ecommerce site, no ecommerce features in Google Analytics have ever been taken advantage of. So I have no way to determine what the actual value of a page is. I've been tasked with determining if a particular page on the site is worth keeping around. How might I use all standard data (pageviews, bounce rate, time on page, time on site, etc.) to help determine the value of this page? I really appreciate any help I can get!

    Read the article

  • Ubuntu 12.04 LTS : la beta 2 de "Precise Pangolin" améliore HUD, Ubuntu One et intègre de nouvelles versions d'applications

    Ubuntu 12.04 LTS : la beta 2 de "Precise Pangolin" améliore HUD Ubuntu One et intègre de nouvelles versions d'applications Mise à jour du 09/04/2012 À quelques semaines de la publication de la version finale de Precise Pangolin, la prochaine version du système d'exploitation fondé sur Linux, Canonical publie la dernière beta de l'OS. La beta 2 d'Ubuntu 12.04 LTS apporte des corrections de bugs et quelques nouvelles fonctionnalités, dont l'ajout du noyau Linux 3.2.0-20.33 qui est basé sur la version stable 3.2.12 du Kernel. HUD, le nouveau menu intelligent pour les applications Unity dont un premier aperç...

    Read the article

  • partitioning ssd +hdd

    - by ALdaperan
    I recently bought an ssd drive 128 gb and i would like yo have your suggestions on partitiong and installing ubuntu 12.10 on it. First of all my drives are : HDD 640 Gb SSD Samsung 830 series 128 Gb Whats the best partitiong for theese drives ? 128 Gb is sure enormous amount of space only for / (root) . Is it a good choice if i make 2 partitions on ssd (20 Gb / and 100 /home) and leave hdd as backub drive for my data (files,movies,music etc) ? In this case what mount point must have hdd ? Can you suggest me the best partitiong for my drives ?

    Read the article

  • Finding the try for an except or finally [migrated]

    - by ?s?
    I'm dealing with some code that has fantastically long methods (10k lines!) and some odd use of try-finally and try-except blocks. Some of the latter are long by themselves, and don't always have the try at the start of the method. Obviously I'm trying to refactor the code, but in the meantime just being able to fix a couple of common pathologies would be much easier if I could jump to the start of a block and see what is happening there. When it's 20+ pages away finding it even with the CNPack rainbows is just tedious. I'm using D2010 and have GExperts (with DelForExp), CNPack and DDevExtensions installed, but I can't find anything that lets me jump from the try to the finally or back. Am I missing something? Is there another add-in that I can use that will get me this?

    Read the article

  • Configuring MySQL Cluster Data Nodes

    - by Mat Keep
    0 0 1 692 3948 Homework 32 9 4631 14.0 Normal 0 false false false EN-US JA X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:12.0pt; font-family:Cambria; mso-ascii-font-family:Cambria; mso-ascii-theme-font:minor-latin; mso-hansi-font-family:Cambria; mso-hansi-theme-font:minor-latin; mso-ansi-language:EN-US;} In my previous blog post, I discussed the enhanced performance and scalability delivered by extensions to the multi-threaded data nodes in MySQL Cluster 7.2. In this post, I’ll share best practices on the configuration of data nodes to achieve optimum performance on the latest generations of multi-core, multi-thread CPU designs. Configuring the Data Nodes The configuration of data node threads can be managed in two ways via the config.ini file: - Simply set MaxNoOfExecutionThreads to the appropriate number of threads to be run in the data node, based on the number of threads presented by the processors used in the host or VM. - Use the new ThreadConfig variable that enables users to configure both the number of each thread type to use and also which CPUs to bind them too. The flexible configuration afforded by the multi-threaded data node enhancements means that it is possible to optimise data nodes to use anything from a single CPU/thread up to a 48 CPU/thread server. Co-locating the MySQL Server with a single data node can fully utilize servers with 64 – 80 CPU/threads. It is also possible to co-locate multiple data nodes per server, but this is now only required for very large servers with 4+ CPU sockets dense multi-core processors. 24 Threads and Beyond! An example of how to make best use of a 24 CPU/thread server box is to configure the following: - 8 ldm threads - 4 tc threads - 3 recv threads - 3 send threads - 1 rep thread for asynchronous replication. Each of those threads should be bound to a CPU. It is possible to bind the main thread (schema management domain) and the IO threads to the same CPU in most installations. In the configuration above, we have bound threads to 20 different CPUs. We should also protect these 20 CPUs from interrupts by using the IRQBALANCE_BANNED_CPUS configuration variable in /etc/sysconfig/irqbalance and setting it to 0x0FFFFF. The reason for doing this is that MySQL Cluster generates a lot of interrupt and OS kernel processing, and so it is recommended to separate activity across CPUs to ensure conflicts with the MySQL Cluster threads are eliminated. When booting a Linux kernel it is also possible to provide an option isolcpus=0-19 in grub.conf. The result is that the Linux scheduler won't use these CPUs for any task. Only by using CPU affinity syscalls can a process be made to run on those CPUs. By using this approach, together with binding MySQL Cluster threads to specific CPUs and banning CPUs IRQ processing on these tasks, a very stable performance environment is created for a MySQL Cluster data node. On a 32 CPU/Thread server: - Increase the number of ldm threads to 12 - Increase tc threads to 6 - Provide 2 more CPUs for the OS and interrupts. - The number of send and receive threads should, in most cases, still be sufficient. On a 40 CPU/Thread server, increase ldm threads to 16, tc threads to 8 and increment send and receive threads to 4. On a 48 CPU/Thread server it is possible to optimize further by using: - 12 tc threads - 2 more CPUs for the OS and interrupts - Avoid using IO threads and main thread on same CPU - Add 1 more receive thread. Summary As both this and the previous post seek to demonstrate, the multi-threaded data node extensions not only serve to increase performance of MySQL Cluster, they also enable users to achieve significantly improved levels of utilization from current and future generations of massively multi-core, multi-thread processor designs. A big thanks to Mikael Ronstrom, Senior MySQL Architect at Oracle, for his work in developing these enhancements and best practices. You can download MySQL Cluster 7.2 today and try out all of these enhancements. The Getting Started guides are an invaluable aid to quickly building a Proof of Concept Don’t forget to check out the MySQL Cluster 7.2 New Features whitepaper to discover everything that is new in the latest GA release

    Read the article

  • sql server 2008 insert statement question

    - by user61752
    I am learning sql server 2008 t-sql. To insert a varchar type, I just need to insert a string 'abc', but for nvarchar type, I need to add N in front (N'abc'). I have a table employee, it has 2 fields, firstname and lastname, they are both nvarchar(20). insert into employee values('abc', 'def'); I test it, it works, seems like N is not required. Why we need to add N in front for nvarchar type, what's the pro or con if we are not using it?

    Read the article

  • Sites with overlapping code-bases. Developing multiple sites with little changes

    - by Web Developer
    I have to develop 3 different sites video.com for hosting video audio.com for hosting audio docs.com for hosting docs. domain names for example only Almost 80% of the functionality is the same for all the three, with remaining 20% being completely different features... How do I handle this? How does sites like SO handle this? I am developing this in YII framework and was thinking of having these different features as modules but in this case the menu/code links in html code can become difficult.

    Read the article

  • Slow writing HDD speed, Ubuntu 12.04 64-bit, Thinkpad T520i

    - by pyc
    It seems that (but I'm not completely sure), that when I'm copying files from gigabit network to HDD, I can't use full potential of the network which in my case is about 60 MB/s, because HDD writing is so slow like lower than 10 MB/s, and also it's slowing down the whole system which becomes pretty much unresponsive, almost impossible to work with. Copying files to samba share residing at Ubuntu machine, connected to share from Windows 7, I'm completely sure my network equipment is OK, and there's no CPU intensive process on Ubuntu except smbd getting about 10-20% from time to time which I think is OK. Something here is burried deep I think, maybe even in kernel. Already tried to switch from AHCI to compatibility mode, and turning acpi on and off - nothing helped. So it's like HDD buffer is full and emptying slowly while machine is sluggish, load is about 3 to 4. Somebody experienced the similar problems? Some help on troubleshooting process and identifying the cause would be helpful too :) Thanks!

    Read the article

  • Modelio passe en version 2.1.1, l'outil de modélisation augmente son ouverture à l'open source

    Modelio passe en version 2.1.1 L'outil de modélisation augmente son ouverture à l'open source Modelio, l'outil de modélisation pour le développement de logiciels, la gestion de processus métiers et l'ingénierie des systèmes, passe en version 2.1.1, et accentue son ouverture open source. Modelio est par exemple désormais disponible nativement en format 64 bits ou 32 bits sous les différentes plateformes Linux RedHat, Ubuntu et Debian et gère la documentation Libre Office (en plus du HTML et de Microsoft Word). Résultat de plus de 20 ans de développement propriétaire, l'environnement open source Modelio 2 est disponible sous licence GPL v2 et est doté d'une architecture modulaire, d...

    Read the article

< Previous Page | 148 149 150 151 152 153 154 155 156 157 158 159  | Next Page >