Search Results

Search found 7671 results on 307 pages for 'slow browsing'.

Page 146/307 | < Previous Page | 142 143 144 145 146 147 148 149 150 151 152 153  | Next Page >

  • Reading the memory of a N64

    - by toazron1
    I'm looking for a way to read the memory of a N64, while the game is running, in real time. I have a c# program which hooks into the memory of a running emulator and tracks SSB64 stats. I want to do the same thing with the physical N64. Currently it is possible to read the memory with a gameshark pro, however it's extremely slow, buggy, and not practical for what I am trying to accomplish. Would it be possible to tap into the gameshark, or the N64 directly, to access the memory in real time? Thanks!

    Read the article

  • accessing webpages from terminal

    - by August
    Using Ubuntu 12.04 . I know two methods to access the web-pages through terminal . They are lynx & elinkswhen i have tried to use them to access web-pages from terminal i didn't felt any kind of speed improvements . I am not sure whether its a usual thing or my speed (because its slow one ). so which way is the best to access the web-page terminal browsers or GUI browsers? and what else more i can get/lost from text terminals ?

    Read the article

  • Does anyone use 12.04 with HD 6950?

    - by Midori
    Back when I first tried to switch from Windows to Ubuntu (6-7 months ago) I had several graphics related problems. The VGA was overheating, graphic glitches, video playing issues, slow/choppy window movements etc. After 2-3 reinstalls, hours of playing with different drivers/settings I decided to return to Windows :\ Now I want to give it another shot. But before I start to format, partitioning, backup and other time consuming stuff I wanted to ask if anyone using the 12.04 with HD 6950. Can I utilize the full potential of it, or the drivers are still not good enough? I know that Linux isn't meant to play games in the first place, but the games I playing (SC2, BLC, HoN, DotA 2) are working with wine as far as I know (or aren't?) and I can't find any reason not to switch from Windows if I can utilize the full potential of my config in Ubuntu. So anyone who got experiences with this VGA in Ubuntu please reply. Thanks in advance :)

    Read the article

  • Buy the server contents

    - by user103475
    How i can buy the contents of the server of archive.ubuntu.com (for lucid lynx)? In my country i have ultra slow connection. It will take years for downloading all these gigabytes. Because, canonical will stop supporting ubuntu 10.04, i want all the files now, before it's too late. i know that all files will be transfer to old.dir but i think some will be lost. And anyway i want the files. My question is: Can canonical provide (with payment) a disk with the contents of the server? Cheers

    Read the article

  • Hardware question; booting ubuntu off of USB flash drive?

    - by Kon
    I'm currently running ubuntu off of a USB 3.0 flash drive on my laptop, and I have another USB 3.0 flash drive I use for storing larger files that I need to access and whatnot. But the issue is that both of the usb 3.0 flash drives don't fit next to each other at all, so I'm forced to use another usb 2.0 port located in a different spot on my laptop. Now my question is that if I use a USB 3.0 hub will that slow my connections down at all? or cause some sort of interference? Another question I had which I was unable to find anywhere on the internet is if there was a USB 3.0 hub with more than one upstream connection? Thanks for your time!

    Read the article

  • b43 Firmware Error when trying to install Ubuuntu 14.04 on XP Laptop

    - by WizardNo.7
    I know I'm an Ubuntu and Linux n00b, but I've been browsing the web for a few days searching for a solution, and none of the other b43 posts etc solve my problem. I am trying to install Ubuntu 14.04 on a Fujitsu Siemens Amilo Pro laptop(It's quite old yes, has about 30GB Hard Drive and I think 192mb of RAM) which currently has Windows XP installed(which I'd like to keep for the time being). I have downloaded the 32-bit Desktop ISO and used unetbootin to create a Live USB for this laptop. When I boot from USB, I arrive at the unetbootin Grey and Blue menu and pick either "Try Ubuntu without installing", or "Install Ubuntu". The menu goes away and an Ubuntu loadscreen showing UBUNTU and four dots which progressively change between white and orange. At about the second color changing cycle a white underscore symbol appears next to the fourth dot and flickers, and then I get the usual: [ 95.514833] b43-phy0 ERROR: Firmware file "b43/ucode5.fw" not found [ 95.514847] b43-phy0 ERROR: Firmware file "b43-open/ucode5.fw" not found [ 95.514855] b43-phy0 ERROR: You must go to http://wireless.kernel.org/en/users/Drivers/b43#devicefirmware and download the correct firmware for this driver version. Please carefully read all the instruction on this website. After this I have to hard reboot or shut-down. I don't have a previous Linux install so cannot do the apt-get from my previous install. I cannot access the boot options menu by using F6 or similar. Any ideas? Thanks beforehand!

    Read the article

  • Problem with loading webpage in firefox

    - by user107735
    I'm using ubuntu 12.04 and I always keep my pc up to date. For last 4 or 5 weeks, I noticed in firefox, the page displays enough data but spinner in the tab doesn't stop. Not only that, Some page load very quickly and wheel still speaning. Some are very very slow and wheel is as always speaning. Some websites (ex. google analytics) takes more than 5 minutes to load. Some websites load but slider doesnt load and wheel is speaning (ex. blenderdesh.com) I'm facing lot of problems for this. Please help me.....

    Read the article

  • how to serialize function depending on what instance of object calls it, if same instance call in a thread then do serialize else not

    - by LondonDreams
    I have a function which fetches and updates some record from db and I am trying to make sure each if the function is called by same instance of object(same Or different thread) then function should behave synchronized else its a call from different object instance function need not to be synchronized. I have tried it use a lock per client. That is, instead of synchronizing the method directly using explicit locking through lock objects using Map. function is like :- getAndUpdateMyHitCount(myObjId){ //go to db and get unique record by myObjId //fetch value , increment , save update } And this function may get call is same thread by different Or same object instance But as fetching and matching from Map is slow , Is there other optimized way to do this ? Found similar at this Question but dont feel that is optimized

    Read the article

  • Continuing to code on large projects

    - by user3487347
    I am a hobbyist programmer, and I've started many medium - sized projects to work on just by myself. These include games, a raytracer, physics simulations etc. By the time these projects get to a certain size (around 5000 lines), I begin to slow down in adding features to the program. This is not because of a lack of ideas of what to implement in a program, but rather a struggle in how to go about it. In particular, I'm afraid of breaking what I already have working in order to implement a new feature. I've tried using version control like Git and Subversion, but these seem unnecessary when you are a one man team. I simply have a folder of "versions" of my program, one for each major change I make. How do I keep coding past this 5000 line mark? What about the 50000 line mark?

    Read the article

  • Release Notes for 3/2/2012

    Here are the notes for today’s release: Added a progress indicator when saving issues. Added support for viewing CodePlex RSS feeds in Chrome. Deployed several bug fixes: Fixed an issue where the back button on Internet Explorer was not working as intended when browsing code. Fixed an issue where long commit comments would push the source control info box outside of the boundaries of the page. Fixed an issue where Internet Explorer users were not able to widen the frame of the source code browser until a file was selected. Fixed an issue where opening a source code file directly from a URL in Internet Explorer would cause the source code tree to be collapsed. Fixed an issue where adding a code snippet with long lines of text to a discussion thread using Internet Explorer would needlessly display a vertical scrollbar, limiting the amount of code visible. Fixed an issue where tabbing through some links would render them invisible. We deprecated support for embedding PreEmptive analytics statistics on the project statistics page. If you’re interested in collecting and reporting your own statistics, PreEmptive’s RunTime Intelligence Endpoint Starter Kit offers a good starting point for capturing data. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

  • Internet Not Working... Initially Worked fine

    - by Sneha429
    I am using an HP Pavilion desktop, not sure of exact model, etc. Initially saucy looked and worked great. It's my parents, and after two days of no issues, they started having more than a few. The mostly use the internet to access email and videos. I believe Google Chrome was the only application they used. Then, the internet stopped working. Restarting didn't fix it, neither did turning the internet on and off. The computer went from lightening fast to inexplicably slow. I've read some other posts about the internet not working, but none seemed applicable. Any suggestions?

    Read the article

  • System doesn't boot due to a "[drm: drm_crtc_helper_set_config] * error * "

    - by Albin
    I have a problem with the new Ubuntu 12.10. It finished installing correctly, but when I restart my computer an error message appears: [15.854296] [drm: drm_crtc_helper_set_config] * error * failed to set mode on [CRTC: 10] * checking battery state [17.814209] [drm: drm_crtc_helper_set_config] * error * failed to set mode on [CRTC: 10] and that screen does not advance. I also tried booting from the recovery mode and it starts, but it's too slow loading windows and also shows me a window with an error message: Sorry Ubuntu 12.10 has experienced an internal error More Details: Executable path /usr/bin/xorg I have used Ubuntu since the 2008 version and never had this problem. I hope you can help me. Computer Specs: compaq sr2015la AMD sempron 2Gb. RAM 80Gb Hard Drive Translated from the original Spanish

    Read the article

  • Only can move the pointer after upgrading

    - by Mystogan
    I made a partition on my desktop pc, 100 GB and installed Ubuntu 12.10 on it. It was pretty slow first thought it had to install updates or something, so I rebooted it. Now after the login it looks like everything freezes, I can move my mouse only, but that's it! I can see my desktop and launcher, but the task bar doesn't display time and misc system info, only a black bar, I downloaded the Android SDK pack and installed vlc media player before I rebooted. My computer info: 1 TB HDD 8 GB ddr3 RAM ATI Radeon 4890 videocard (1 GB) AMD Phenom 2 x4 black 3, 2ghz CPU (64 bits)

    Read the article

  • Can it be a good idea to create a new table for each client of a webapp?

    - by Will
    This is semi-hypothetical, and as I've no experience in dealing with massive database tables, I have no idea if this is horrible for some reason. On to the situation: Imagine a web based application - lets say accounting software - which has 20,000 clients and each client has 1000+ entries in a table. That's 20 million rows which I know can certainly slow down complex queries. In a case like this, does it make more sense to create a new table in the database for each client? How do databases react to having 20k (or more!) tables?

    Read the article

  • Is this JS code a good way for defining class with private methods?

    - by tigrou
    I was recently browsing a open source JavaScript project. The project is a straight port from another project in C language. It mostly use static methods, packed together in classes. Most classes are implemented using this pattern : Foo = (function () { var privateField = "bar"; var publicField = "bar";     function publicMethod() { console.log('this is public');     } function privateMethod() { console.log('this is private'); } return {   publicMethod : publicMethod, publicField : publicField }; })(); This was the first time I saw private methods implemented that way. I perfectly understand how it works, using a anonymous method. Here is my question : is this pattern a good practice ? What are the actual limitations or caveats ? Usually i declare my JavaScript classes like that : Foo = new function () { var privateField = "test"; this.publicField = "test";     this.publicMethod = function()     { console.log('this method is public'); privateMethod();     } function privateMethod() { console.log('this method is private'); } }; Other than syntax, is there any difference with the pattern show above ?

    Read the article

  • How to make Ubuntu 14.04 run with less lag?

    - by King Shimkus
    I recently updated from Ubuntu 12.04 LTS to 14.04 LTS and for the first time it worked! However, whenever I move my mouse over an application icon on the unity shell, the animation takes forever to show me what the name of the application is ( A.k.a the tooltip ). The same happens with menus and sub-menus. Other than that, it is overall in general, slow. I just want to know if there are any tips to fix this or make my system faster. This is what it says when I type in glxinfo | grep renderer: GLX_MESA_multithread_makecurrent, GLX_MESA_query_renderer, OpenGL renderer string: Gallium 0.4 on llvmpipe (LLVM 3.4, 128 bits) Output of lspci | grep VGA: 00:02.0 VGA compatible controller: Intel Corporation 82865G Integrated Graphics Controller (rev 02)

    Read the article

  • Browser Item Caching and URLs

    - by Damon
    Ultimately you want the browser to cache things like Flash components, Silverlight XAP files, and images to avoid users having to download them each time they hit a page.  But during development it's very useful to NOT have things cached so you are always looking at the most up-to-date file.  You can always turn off caching on your browser, but if you use your browser for daily browsing then its not the greatest option.  To avoid caching we would always just slap a randomly generated GUID to the back of the URL of any items we didn't want to cache (e.g. http://someserver.com/images/image.png?15f073f5-45fc-47b2-993b-fbaa781b926d).  It worked well, but you had to remember to remove the random GUID when it went to production. However, on a GimmalSoft project we recently implemented someone showed me a better way that didn't need to be removed from production code - just slap the last modified date of the file on the end of the URL (or something generated from the modification date).  This was kind of genius approach because it gives you the best of both world.  If you modify the file, the browser goes out and gets the newest version.  If you don't modify the file, it has the cached copy.  Very helpful!  The only down side is that you do have to read the modification date from the file, which does technically take some time.

    Read the article

  • Browser Item Caching and URLs

    - by Damon Armstrong
    Ultimately you want the browser to cache things like Flash components, Silverlight XAP files, and images to avoid users having to download them each time they hit a page.  But during development it’s very useful to NOT have things cached so you are always looking at the most up-to-date file.  You can always turn off caching on your browser, but if you use your browser for daily browsing then its not the greatest option.  To avoid caching we would always just slap a randomly generated GUID to the back of the URL of any items we didn’t want to cache (e.g. http://someserver.com/images/image.png?15f073f5-45fc-47b2-993b-fbaa781b926d).  It worked well, but you had to remember to remove the random GUID when it went to production. However, on a GimmalSoft project we recently implemented someone showed me a better way that didn’t need to be removed from production code – just slap the last modified date of the file on the end of the URL (or something generated from the modification date).  This was kind of genius approach because it gives you the best of both world.  If you modify the file, the browser goes out and gets the newest version.  If you don’t modify the file, it has the cached copy.  Very helpful!  The only down side is that you do have to read the modification date from the file, which does technically take some time.

    Read the article

  • Best way to get elapsed time in miliseconds in windows

    - by XaitormanX
    I'm trying to do it using two FILETIMEs, casting them to ULONGLONGs, substracting the ULONGLONGs, and dividing the result by 10000. But it's pretty slow, and I want to know if there is a better way to do it.I use c++ with visual studio 2008 express edition. This is what I'm using: FILETIME filetime,filetime2; GetSystemTimeAsFileTime(&filetime); Sleep(100); GetSystemTimeAsFileTime(&filetime2); ULONGLONG time1,time2; time1 = (((ULONGLONG) filetime.dwHighDateTime) << 32) + filetime.dwLowDateTime; time2 = (((ULONGLONG) filetime2.dwHighDateTime) << 32) + filetime2.dwLowDateTime; printf("ELAPSED TIME IN MS:%d",(int)((time2-time1)/10000));

    Read the article

  • Why does my flash video get stuck? (not lag?)

    - by Sk606
    After moving to Ubuntu 10.10 Netbook Edition, I've noticed a strange problem that manifests itself most obviously with streaming videos like youtube and hulu. The video will load and begin playing fine and eventually stop, as though it is buffering. However, it usually happens between 40 and 60 seconds into the clip. The loading indication however shows that it actually isn't buffering anymore of the data, and will just sit there. If I force the clip forward (e.g. click to start the movie somewhere ahead of the stall point) the clip will usually reload quickly and begin playing as expected. Because of the quick load times, and the consistent stop of video, I don't think the problem is related to hardware performance or network throughput. Also, I have none of these problems in 10.04. Rebooting into 10.04 makes the problem disappear. However, one more clue, I also have noticed a higher occurrence of web pages that simply "stop loading" - say, 1/10 clicks while browsing I have to hit "reload" to get the page to load because the page timed out part of the way through. This seems strangely related to the video problem, though is a lot less infuriating. Help! Where do I begin to look with a problem like this?

    Read the article

  • How can I run a complete application from RAM?

    - by pico
    When I do some stuff in background (unpacking, compiling, backup, etc.) my hdd is under load and for example, Firefox and Chromium take very long to start and react very slow. While my hdd is under load I have still plenty of free RAM and free CPU. How can I copy the whole Firefox or Chromium including all dependencies into RAM? I don't care about persistence, I just want to view some websites, videos, etc. There are instructions putting the cache into /dev/shm, but that didn't speed up so much. Getting the profile into the RAM drive might also be easy... But how to get the binaries with all dependencies into RAM?

    Read the article

  • Getting expired domain name - most effective route?

    - by kcdwayne
    There is a domain name I have been wanting for years that was used as a parked page (read: cybersquatted) that has entered into WHOIS's redemptionPeriod stage. The domain has been expired for 61 days now - after contacting the Registar, they informed me that it would stay in redemption for 75 days. After which, it would either be sold to resellers, or sent back to the public registry. (I have since sent a followup message requesting the reseller they use.) My question is: What is the best way to proceed? I know there is at least one competitor that would love to have this name, but I'm unsure if they even know it's expiring. I did not tell the Registar the domain in question, as they seem geared towards cybersquatting, and I do not trust them. Domain Front Running sucks. Should I use a backorder service? Should I just take my chances and try to grab it after75 days? I checked an auctionhouse by manually browsing their expired domains - it wasn't there.

    Read the article

  • Intersection of player and mesh

    - by Will
    I have a 3D scene, and a player that can move about in it. In a time-step the player can move from point A to point B. The player should follow the terrain height but slow going up cliffs and then fall back, or stop when jumping and hitting a wall and so on. In my first prototype I determine the Y at the player's centre's X,Z by intersecting a ray with every triangle in the scene. I am not checking their path, but rather just sampling their end-point for each tick. Despite this being Javascript, it works acceptably performance-wise. However, because I am modeling the player as a single point, the player can position themselves so that they are half-in a cliff face and so on. I need to model them as as a solid e.g. some cluster of spheres or a even their fuller mesh. I am also concerned that if they were moving faster they might miss the test altogether. How should I solve this?

    Read the article

  • Ubuntu One lost most of my files!

    - by Max
    I keep in my UbuntuOne the source code of my projects as a back up. In the past it worked and I never had problem. It happened that I had to change the hard disk of my laptop and I installed a fresh Ubuntu 12.10 in the new one. First time I connected with UbuntuOne it downloaded my files, but when I went to see my projects almost all my c++ source code files were missing!!!... I tried to check if I still have them on UbuntuOne by accessing from the web but nothing... my work is lost forever. I don't know who to ask for help? Is there a way to get back my precious files? Honestly I can't trust this service any more, I'm very disappointed. (edit: thanks God I found a back up in one of my external hard disks. I won't trust Ubuntu One any more, it's buggy and quite slow. )

    Read the article

  • Images from remote source - is it possible or is really a bad practice?

    - by user1620696
    I'm building a management system for websites and I had an idea related to image galleries that I'm not sure it's a good approach. Since images might need good deals of space depending on how much images a user uploads an so on, I thought on using cloud services like dropbox, mega and google drive to store images and load then when needed. The obvious problem is that for me this seems a useless solution because it would be slow to download the images from the remote source, making the user experience not so good. Is there any way to save images of a image gallery on remote source without getting the user experience bad because of speed? Or this is really not a good practice?

    Read the article

< Previous Page | 142 143 144 145 146 147 148 149 150 151 152 153  | Next Page >