Search Results

Search found 3661 results on 147 pages for 'overload resolution'.

Page 76/147 | < Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >

  • # id - urls with id first display full page, then move to #id

    - by guisasso
    I've noticed this in the new version of chrome, and ie9 and 10. Some urls in a photo gallery have a #id tag as they are supposed to display a full view of a picture. Basically, a div in a lower position on the page has that #id that i call via a.com/1.html#id. This has never been an issue until lately, when i noticed a bit of a lag. The issue: The website loads normally, then the view moves to the #id as supposed, but with some lag sometimes, perhaps because of the high resolution of the picture, which is somewhat noticeable. Anyway to avoid this, or make it so the page would move to the correct #id even before fully loaded?

    Read the article

  • When creating an library published on CodePlex, how "bad" would it be for the unit-test projects to rely on commercial products?

    - by Lasse V. Karlsen
    I have started a project on CodePlex for a WebDAV server implementation for .NET, so that I can host a WebDAV server in my own programs. This is both a learning/research project (WebDAV + server portion) as well as a project I think I can have much fun with, both in terms of making it and using it. However, I see a need to do mocking of types here in order to unit-testing properly. For instance, I will be relying on HttpListener for the web server portion of the WebDAV server, and since this type has no interface, and is sealed, I cannot easily make mocks or stubs out of it. Unless I use something like TypeMock. So if I used TypeMock in the unit-test projects on this library, how bad would this be for potential users? The projects are made in C# 3.5 for .NET 3.5 and 4.0, and the project files was created with Visual Studio 2010 Professional. The actual class libraries you would end up referencing in your software would of course not be encumbered with anything remotely like this, only the unit-test libraries. What's your thoughts on this? As an example, I have in my old code-base, which is private, the ability to just initiate a WebDAV server with just this: var server = new WebDAVServer(); This constructs, and owns, a HttpListener instance internally, and I would like to verify through unit-tests that if I dispose of this server object, the internal listener is disposed of. If, on the other hand, I use the overload where I hand it a listener object, this object should not be disposed of. Short of exposing the internal listener object to the outside world, something I'm a bit loath to do, how can I in a good way ensure that the object was disposed of? With TypeMock I can mock away parts of this object even though it isn't accessed through interfaces. The alternative would be for me to wrap everything in wrapper classes, where I have complete control.

    Read the article

  • Desktop Fun: Heroes of DC Comics Wallpaper Collection

    - by Asian Angel
    DC Comics has a great line-up of super hero characters and we have gathered together some of DC’s finest for you in today’s collection. Give your desktop super powers with our Heroes of DC Comics Wallpaper collection. Note: Click on the picture to see the full-size image—these wallpapers vary in size so you may need to crop, stretch, or place them on a colored background in order to best match them to your screen’s resolution. How to Make and Install an Electric Outlet in a Cabinet or DeskHow To Recover After Your Email Password Is CompromisedHow to Clean Your Filthy Keyboard in the Dishwasher (Without Ruining it)

    Read the article

  • UDDI vs SO-Aware: Why SO-Aware is the More Efficient and Interoperable Alternative

    - by Vishal
    Hello folks,   If you are implementing a service oriented architecture, and are unsure of the best governance approach to follow, then this webinar is a must-attend event for you.  We will discuss why SO-Aware is the more efficient and interoperable alternative to traditional UDDI-based SOA-governance.   Specifically, we will address the differences between UDDI and SO-Aware in terms of service discovery, configuration, and policy resolution.  Finally, we will address why the REST/Odata based model implemented by SO-Aware enables the most efficient governance not only for WCF but for BizTalk, the Windows Server AppFabric and the Windows Azure AppFabric as well.   Join us on January 26th at 2:00 ET - to register, click here    Thanks,   Vishal

    Read the article

  • Multiple vulnerabilities in Thunderbird

    - by RitwikGhoshal
    CVE DescriptionCVSSv2 Base ScoreComponentProduct and Resolution CVE-2012-1948 Denial of service (DoS) vulnerability 9.3 Thunderbird Solaris 10 SPARC: 145200-12 X86: 145201-12 CVE-2012-1950 Address spoofing vulnerability 6.4 CVE-2012-1951 Resource Management Errors vulnerability 10.0 CVE-2012-1952 Resource Management Errors vulnerability 9.3 CVE-2012-1953 Improper Restriction of Operations within the Bounds of a Memory Buffer vulnerability 9.3 CVE-2012-1954 Resource Management Errors vulnerability 10.0 CVE-2012-1955 Address spoofing vulnerability 6.8 CVE-2012-1957 Improper Neutralization of Input During Web Page Generation ('Cross-site Scripting') vulnerability 4.3 CVE-2012-1958 Resource Management Errors vulnerability 9.3 CVE-2012-1959 Permissions, Privileges, and Access Controls vulnerability 5.0 CVE-2012-1961 Improper Input Validation vulnerability 4.3 CVE-2012-1962 Resource Management Errors vulnerability 10.0 CVE-2012-1963 Permissions, Privileges, and Access Controls vulnerability 4.3 CVE-2012-1964 Clickjacking vulnerability 4.0 CVE-2012-1965 Improper Neutralization of Input During Web Page Generation ('Cross-site Scripting') vulnerability 4.3 CVE-2012-1966 Permissions, Privileges, and Access Controls vulnerability 4.3 CVE-2012-1967 Arbitrary code execution vulnerability 10.0 CVE-2012-1970 Denial of service (DoS) vulnerability 10.0 CVE-2012-1973 Resource Management Errors vulnerability 10.0 CVE-2012-3966 Improper Restriction of Operations within the Bounds of a Memory Buffer vulnerability 10.0 This notification describes vulnerabilities fixed in third-party components that are included in Oracle's product distributions.Information about vulnerabilities affecting Oracle products can be found on Oracle Critical Patch Updates and Security Alerts page.

    Read the article

  • Ubuntu 12.04 "Do nothing" when "lid closed" blanks out external monitor

    - by Elijah Lynn
    This is not the same as this question. I have Ubuntu 12.04 running with an Nvidia card on a W510 Thinkpad. I have one external monitor connected. When I change the power settings to "Do nothing" when "Lid closed" it still keeps the system running which is great. However, it blanks out the display on any external monitors making the system useless. I plan on getting a dock soon and having to identical resolution monitors and would love to be able to dock the monitor and work as normal on the external monitors. Does anyone have a suggestion or fix for this? Should I report this as a bug or feature request?

    Read the article

  • 12.10 Booting Into Variations Of Blank Screen

    - by user93954
    I've been running the Ubuntu 12.10 beta since about a month before the final release with almost no problems. However since the day of release (I'm assuming an update has caused this) I have had problems booting into the actual GUI interface. Trying to get it to work is just a case of hard shutting down until it works, but for most of the time I need to battle with various different kinds of black screens. These include a plain black screen, a flashing line and a flashing line that doesn't fit the resolution. Nine times out of ten the cursor will be displayed over these. It also sometimes manages to boot into Ubuntu, albeit text mode or sometimes it loads GRUB which it isn't setup to do. If anyone could help out with this it'd be great. I really, really don't want to have to go through yet another clean installation. Cheers.

    Read the article

  • 12.04 Dual Monitor Configuration changes to "Mirrored" automatically

    - by Josef
    After installing 12.04 on my Samsung X120 Notebook (with Intel GMA 4500 internal graphics card) i have a Problem with the configuration off my 2nd monitor. Its a 22" with 1680 * 1050 Pixel connected via HDMI. I use the display configuration and set internal display=off and external display=main display. That works fine. After rebooting i still have that configuration, but after a while (often when i press a button on the keyboard), the configuration automatically changes to "Mirror Screen" and the internal display switches on. Both displays then have a bad resolution. When i do the manual display configuration again, everything is fine for the duration of that session. What could be the problem, that this display configuration gets lost / resets?

    Read the article

  • Asus Eee PC 701 4G hangs on boot

    - by Andrew
    I've got an old Eee PC 701 4G with the following specifications: 512 MB RAM 4GB SSD drive SM223AC 8GB SD card extension Screen resolution: 800 x 480 BIOS Revision 1101 (05/16/2008) EC Firmware version: EPC-079 Windows XP SP3 works fine on it, but I decided to switch my OS to Ubuntu. I have downloaded an Ubuntu 10.10 Netbook Remix ISO and wrote it to my FAT32 SD card using Universal-USB-Installer-1.8.3.3, as described on ubuntu.com During standard load from the SD card the boot process hangs up with black screen. If I'll press F6 while preloading Ubuntu, it sucessfully displays the boot menu, selecting language and showing 2 main commands: "Run ubuntu from USB drive" and "Install Ubuntu". Selecting either of these commands leading to the same result - after some background work the main loading indicator is displayed ("Ubuntu" text with dotted progress bar under it), and it's progressing forever without any effect. Is Ubuntu 10.10 compatible with my Eee PC at all? How to boot it correctly?

    Read the article

  • Matrix Multiplication with C++ AMP

    - by Daniel Moth
    As part of our API tour of C++ AMP, we looked recently at parallel_for_each. I ended that post by saying we would revisit parallel_for_each after introducing array and array_view. Now is the time, so this is part 2 of parallel_for_each, and also a post that brings together everything we've seen until now. The code for serial and accelerated Consider a naïve (or brute force) serial implementation of matrix multiplication  0: void MatrixMultiplySerial(std::vector<float>& vC, const std::vector<float>& vA, const std::vector<float>& vB, int M, int N, int W) 1: { 2: for (int row = 0; row < M; row++) 3: { 4: for (int col = 0; col < N; col++) 5: { 6: float sum = 0.0f; 7: for(int i = 0; i < W; i++) 8: sum += vA[row * W + i] * vB[i * N + col]; 9: vC[row * N + col] = sum; 10: } 11: } 12: } We notice that each loop iteration is independent from each other and so can be parallelized. If in addition we have really large amounts of data, then this is a good candidate to offload to an accelerator. First, I'll just show you an example of what that code may look like with C++ AMP, and then we'll analyze it. It is assumed that you included at the top of your file #include <amp.h> 13: void MatrixMultiplySimple(std::vector<float>& vC, const std::vector<float>& vA, const std::vector<float>& vB, int M, int N, int W) 14: { 15: concurrency::array_view<const float,2> a(M, W, vA); 16: concurrency::array_view<const float,2> b(W, N, vB); 17: concurrency::array_view<concurrency::writeonly<float>,2> c(M, N, vC); 18: concurrency::parallel_for_each(c.grid, 19: [=](concurrency::index<2> idx) restrict(direct3d) { 20: int row = idx[0]; int col = idx[1]; 21: float sum = 0.0f; 22: for(int i = 0; i < W; i++) 23: sum += a(row, i) * b(i, col); 24: c[idx] = sum; 25: }); 26: } First a visual comparison, just for fun: The beginning and end is the same, i.e. lines 0,1,12 are identical to lines 13,14,26. The double nested loop (lines 2,3,4,5 and 10,11) has been transformed into a parallel_for_each call (18,19,20 and 25). The core algorithm (lines 6,7,8,9) is essentially the same (lines 21,22,23,24). We have extra lines in the C++ AMP version (15,16,17). Now let's dig in deeper. Using array_view and extent When we decided to convert this function to run on an accelerator, we knew we couldn't use the std::vector objects in the restrict(direct3d) function. So we had a choice of copying the data to the the concurrency::array<T,N> object, or wrapping the vector container (and hence its data) with a concurrency::array_view<T,N> object from amp.h – here we used the latter (lines 15,16,17). Now we can access the same data through the array_view objects (a and b) instead of the vector objects (vA and vB), and the added benefit is that we can capture the array_view objects in the lambda (lines 19-25) that we pass to the parallel_for_each call (line 18) and the data will get copied on demand for us to the accelerator. Note that line 15 (and ditto for 16 and 17) could have been written as two lines instead of one: extent<2> e(M, W); array_view<const float, 2> a(e, vA); In other words, we could have explicitly created the extent object instead of letting the array_view create it for us under the covers through the constructor overload we chose. The benefit of the extent object in this instance is that we can express that the data is indeed two dimensional, i.e a matrix. When we were using a vector object we could not do that, and instead we had to track via additional unrelated variables the dimensions of the matrix (i.e. with the integers M and W) – aren't you loving C++ AMP already? Note that the const before the float when creating a and b, will result in the underling data only being copied to the accelerator and not be copied back – a nice optimization. A similar thing is happening on line 17 when creating array_view c, where we have indicated that we do not need to copy the data to the accelerator, only copy it back. The kernel dispatch On line 18 we make the call to the C++ AMP entry point (parallel_for_each) to invoke our parallel loop or, as some may say, dispatch our kernel. The first argument we need to pass describes how many threads we want for this computation. For this algorithm we decided that we want exactly the same number of threads as the number of elements in the output matrix, i.e. in array_view c which will eventually update the vector vC. So each thread will compute exactly one result. Since the elements in c are organized in a 2-dimensional manner we can organize our threads in a two-dimensional manner too. We don't have to think too much about how to create the first argument (a grid) since the array_view object helpfully exposes that as a property. Note that instead of c.grid we could have written grid<2>(c.extent) or grid<2>(extent<2>(M, N)) – the result is the same in that we have specified M*N threads to execute our lambda. The second argument is a restrict(direct3d) lambda that accepts an index object. Since we elected to use a two-dimensional extent as the first argument of parallel_for_each, the index will also be two-dimensional and as covered in the previous posts it represents the thread ID, which in our case maps perfectly to the index of each element in the resulting array_view. The kernel itself The lambda body (lines 20-24), or as some may say, the kernel, is the code that will actually execute on the accelerator. It will be called by M*N threads and we can use those threads to index into the two input array_views (a,b) and write results into the output array_view ( c ). The four lines (21-24) are essentially identical to the four lines of the serial algorithm (6-9). The only difference is how we index into a,b,c versus how we index into vA,vB,vC. The code we wrote with C++ AMP is much nicer in its indexing, because the dimensionality is a first class concept, so you don't have to do funny arithmetic calculating the index of where the next row starts, which you have to do when working with vectors directly (since they store all the data in a flat manner). I skipped over describing line 20. Note that we didn't really need to read the two components of the index into temporary local variables. This mostly reflects my personal choice, in some algorithms to break down the index into local variables with names that make sense for the algorithm, i.e. in this case row and col. In other cases it may i,j,k or x,y,z, or M,N or whatever. Also note that we could have written line 24 as: c(idx[0], idx[1])=sum  or  c(row, col)=sum instead of the simpler c[idx]=sum Targeting a specific accelerator Imagine that we had more than one hardware accelerator on a system and we wanted to pick a specific one to execute this parallel loop on. So there would be some code like this anywhere before line 18: vector<accelerator> accs = MyFunctionThatChoosesSuitableAccelerators(); accelerator acc = accs[0]; …and then we would modify line 18 so we would be calling another overload of parallel_for_each that accepts an accelerator_view as the first argument, so it would become: concurrency::parallel_for_each(acc.default_view, c.grid, ...and the rest of your code remains the same… how simple is that? Comments about this post by Daniel Moth welcome at the original blog.

    Read the article

  • How do engines avoid "Phase Lock" (multiple objects in same location) in a Physics Engine?

    - by C0M37
    Let me explain Phase Lock first: When two objects of non zero mass occupy the same space but have zero energy (no velocity). Do they bump forever with zero velocity resolution vectors or do they just stay locked together until an outside force interacts? In my home brewed engine, I realized that if I loaded a character into a tree and moved them, they would signal a collision and hop back to their original spot. I suppose I could fix this by implementing impulses in the event of a collision instead of just jumping back to the last spot I was in (my implementation kind of sucks). But while I make my engine more robust, I'm just curious on how most other physics engines handle this case. Do objects that start in the same spot with no movement speed just shoot out from each other in a random direction? Or do they sit there until something happens? Which option is generally the best approach?

    Read the article

  • Star Wars: An Infographic Flowchart

    - by Jason Fitzpatrick
    If you can’t get enough of Star Wars lore, this minimalist set of infographics details major characters, conflicts, and alliances in the Star Wars universe. Courtesy of designer Marc Morera, the series of Star Wars infographics give a quick summary, presents all the major players in the movies, and connects all the players and events via flowchart. Hit up the link below to see all of them in their high-resolution glory. Star Wars Infographic [via Cool Infographics] How To Create a Customized Windows 7 Installation Disc With Integrated Updates How to Get Pro Features in Windows Home Versions with Third Party Tools HTG Explains: Is ReadyBoost Worth Using?

    Read the article

  • I need acpi=off to boot, but this seems to disable wifi

    - by TheExp3rt
    To boot my computer, I have to use acpi=off in kernel options. Doing this apparently causes my wifi adapter to be switched off and prevents my mouse and USB keyboard to not work. If I change acpi=off to nomodeset, it fixes these problems. But then it sets my computer at a lower resolution. I am currently running Kubuntu 13.10, although none of the updates, starting from 12.04, changed the need for acpi=off. I assume I need acpi=off in because I have an oldish computer with an intel motherboard How can I boot without acpi=off? Output of uname -a: 3.11.0-13-generic #20-Ubuntu SMP Wed Oct 23 07:38:26 UTC x86_64 x86_64 x86_64 GNU/Linux

    Read the article

  • Graphics system failure

    - by anonymous coward
    I just upgraded two versions of Ubuntu to Oneiric after a recent set of updates broke Audacious 2.4.something in Maverick for me. When it was all finished my Asus Eee 1215N with the 1366x768 screen and the notoriously unsupported Nvidia Ion2 Optimus chipset was working on minimum compatibility fallbacks at 1024x768 resolution. So I did a little bit of reading (not enough) and tried deleting a couple parts of the graphics system to see if I could force a retry on the autodetection. The result: No graphical shell works now. Attempted so far: -reinstalling xorg/nouveau, didn't work. -installing bumblebee/nvidia-current, didn't work. -sudo dpkg-reconfigure -phigh -a, didn't work. I'd like to stress that this is otherwise a perfectly functional installation that I had spent some time setting up already (including a successful compile of Audacious 3.21 that was working with all the plugins). Is there a way for me to rebuild the graphics system from scratch or should I just give up, cry a little, backup data, wipe partition and reinstall with a USB drive?

    Read the article

  • Joining Two MKV files in Ubuntu?

    - by Ryan McClure
    I have an opera that I'm ripping to my computer in MKV format with Handbrake. This opera is on two discs. Is there a way to join the resulting MKV's together? They will have the same bitrate, resolution, etc. If I do this, can I keep chapters from both MKV files organized? And, since I have subtitles in the file (not burnt in), will they still stay intact? I'm not too sure if this question is off-topic or not. If it is, feel more than free to delete it. :)

    Read the article

  • Narrow-phase collision detection algorithms

    - by Marian Ivanov
    There are three phases of collision detection. Broadphase: It loops between all objecs that can interact, false positives are allowed, if it would speed up the loop. Narrowphase: Determines whether they collide, and sometimes, how, no false positives Resolution: Resolves the collision. The question I'm asking is about the narrowphase. There are multiple algorithms, differing in complexity and accuracy. Hitbox intersection: This is an a-posteriori algorithm, that has the lowest complexity, but also isn't too accurate, Color intersection: Hitbox intersection for each pixel, a-posteriori, pixel-perfect, not accuratee in regards to time, higher complexity Separating axis theorem: This is used more often, accurate for triangles, however, a-posteriori, as it can't find the edge, when taking last frame in account, it's more stable Linear raycasting: A-priori algorithm, useful for semi-realistic-looking physics, finds the intersection point, even more accurate than SAT, but with more complexity Spline interpolation: A-priori, even more accurate than linear rays, even more coplexity. There are probably many more that I've forgot about. The question is, in when is it better to use SAT, when rays, when splines, and whether there is anything better.

    Read the article

  • Virtual screen size with libgdx and GLES 2

    - by David Saltares Márquez
    I've been trying to use a virtual screen size for my libgdx desktop-android game. I'd like to always use a 16:9 aspect ratio but with a virtual screen size so everything would adapt automatically depending on the device size. This post illustrates the process pretty well but my game crashes when camera.apply(Gdx.Gl10) is called. This is because I'm using GLES 2.0 (for not having to use multiple of 2 texture sizes). As stated in the OrthographicCamera doc, the apply method only works with GLES 1 and GLES 1.1. Is there another way of applying my GL transformation to the camera so I can use a virtual screen resolution? Having to resize everything manually it's a total pain. Thanks a lot.

    Read the article

  • OM: Effective Troubleshooting Techniques to Debug Order Import

    - by ChristineS
    There is a new document available to help you debug Order Import in Order Management:  Effective Troubleshooting Techniques to Debug Order Import Issues (Doc ID 1558196.1) The white paper addresses debugging from a technical perspective. This approach is to assist users in understanding the actual issue, as well as help them in the early resolution of any order import issue. It will walk you through several cases, with supporting debug logs / trace files taken for each case. Educating you along the way for what debug logs / trace files should be gathered, as trace files are not always needed.  It will also walk you through the supporting documents so you will know what to look for in your case. Please refer to this note the next time you have an Order Import error. Or you could step through it now, so you are informed the next time you encounter an Order Import error.  Happy debugging!

    Read the article

  • CRT as 2nd monitor goes screwy after start up?

    - by rhys
    new install of 12.04 on an old dell with a radeon ATI RV516 video card with monitor out and s-video out. During boot up all is good. Both screens operate and look fine. Then just before the desktop appears the cCRT goes purple and is covered in heavy horizontal lines, but as i said during boot up it was fine and the resolution was fine. The main monitor, an lcd, operates normally. everything else works fine, it's just the picture on the CRT that is screwed up. I used the same monitor and CRT running 11.10 which worked fine any help would be appreciated and yes i am a newbie to ubuntu here is a vid showing the completely normal screens at reboot then the purple badness when the desktop loads ?? and don't laugh at the slow machine, it's old. http://www.youtube.com/my_videos_edit?ns=1&video_id=zfuh6lBMLnc

    Read the article

  • Apache randomly timing out

    - by Zaid
    I've been wrestling with this problem for few days now. Apache works fine. Then suddenly starts timing out. There is nothing in the error log. Few more things: - I've gone so far as to reinstall the box. - The codebase has not been touched in months. - I've done the speech test so I know it's not a bandwidth overload problem - Restart apache does not necessarily fix the issue, even temporarily (only thing that does is random attempts) If you can guide me to tools that can help me figure this out or if you know any specifics I should see, appreciate it.

    Read the article

  • Lost the config of my monitor once I connected it to a KVM switch.

    - by jfmessier
    I used to have a monitor (Acer 20", AL2017) connected to the external video port of my laptop. Everything was great and I could get a rather high resolution available when setting it under Ubuntu (all versions). But since I connected it to a KVM, the monitor is no more recognized, and I cannot go beyond 1024x768. My video adapter is an Intel chipset 9xx from an older Pentium Core 2 Duo laptop. The same monitor, using another video chipset (Intel 82G33/G31 rev 02) is properly detected through the switch box. Looks to me like the video chipset is not the best, and I can live with this. I see that I have no xorg.conf, and I understand that I would need to generate one so I can then use it to force the available modes on the X system. How can I generate the xorg.conf file for my Intel video chip, so I can use it then ? Merci :-)

    Read the article

  • WP7&ndash;Samsung Owners Should Hold Off on the Update

    - by D'Arcy Lussier
    Microsoft released an update for WP7 that’s meant to improve updating the devices going forward. They’ve identified a glitch though, specifically with Samsung phones. From Winrumors: A number of Windows Phone 7 users applied the patch on Monday and some Samsung Omnia owners devices have been left in a “bricked” state. Devices simply instruct users to connect them to a PC, hard resetting the device or connecting it to a PC does not appear to solve the issue. Microsoft has also been advising users with broken devices to return them to stores for exchange. The scary thing about this is the resolution: return the “broken” devices to stores for exchange. Many Samsung Focus owners in Canada purchased unlocked phones from the US, and supposedly that act alone voided the warranty. So applying the update has very dire implications for those who might be left with a very pretty brick. I’m going to wait until there are successful installs of the patched update before going ahead with it on my device.

    Read the article

  • High Availability documents for OBIA

    - by Lia Nowodworska - Oracle
    There are 2 white papers that have been created by Product Management and Advanced Resolution team (thanks Rajesh, Archana). These documents describe how to deploy a high-availability environment for the Weblogic components of BI Applications 11.1.1.8.1 and 11.1.1.7.1 including the Oracle Data Integrator. New  Configuring High Availability for Oracle Business Intelligence Applications Version 11.1.1.8.1 (Doc ID 1679319.1)  Updated Configuring High Availability for Oracle Business Intelligence Applications Version 11.1.1.7.1 (Doc ID 1587873.1)  When implementing OBIA please take some time to review one or both of the papers. Do you still have a quick question that you want an expert to have a look at: check in the OBIA Community:

    Read the article

  • Half-maximized windows misplaced after switching display

    - by Arild
    I have a laptop which I often plug into a larger screen with another display resolution. Thus I'm frequently changing the screen for the system to use. Also I've recently discovered the advantages of half-maximizing windows when working alternatingly in two or more applications. When changing screens, however the windows basically retain their absolute sizes and positions. So in my case the windows typically have their bottoms outside display, don't fill up half the display, etc. I have to half-maximize the windows again every time I switch screens. I'm thinking this should be subject of a bug report but I find it difficult to know where to file them. Any ideas on what package is responsible?

    Read the article

  • nVidia GeForce Go 7600? can it ever run unity?

    - by Khaled Musleh
    my laptop Toshiba Qosmio G30 has nVidia GeForce Go 7600 card and it suppose to support 3D . i run unity 2d now . I run 12.04 and the graphic driver is--VESA: G73 Board - toshg73m-- by UBUNTU. when i run /usr/lib/nux/unity_support_test -p then i get this list Not software rendered: no Not blacklisted: yes GLX fbconfig: yes GLX texture from pixmap: yes GL npot or rect textures: yes GL vertex program: yes GL fragment program: yes GL vertex buffer object: yes GL framebuffer object: yes GL version is 1.4+: yes Unity 3D supported: no the card is not blacklisted but a similar one with GT is! Do you think that there is a chance the laptop can run the unity 3d? and may be i could change the resolution of the screen to a higher one too! I tried all the nvidia drivers provided but none works (except 96 in ubuntu 12.04 ). i get a black screen or terminal screen. best wishes to all

    Read the article

< Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >