Search Results

Search found 25660 results on 1027 pages for 'booting issue'.

Page 580/1027 | < Previous Page | 576 577 578 579 580 581 582 583 584 585 586 587  | Next Page >

  • Is there a NVIDIA driver (for a 7-series card) that will actually work for 12.10?

    - by DS13
    I see many similar topics on this, but I've tried all their suggestions, and nothing has worked. ISSUE: I do a clean install of Ubuntu 12.10. Boots fine with the “nouveau” graphics driver – graphics are very slow and choppy. The three other driver options in Ubuntu (official NVIDIA drivers), all result in a variation of the black screen on boot up. There will be NO access to a command line/GUI in anyway what-so-ever (tried every option recommended out there, but the system is unusable at this stage). I can only reinstall, and try different drivers…and I only ever get one shot at it. QUESTIONS: Does anyone know of a NVIDIA driver that will actually work with a Nvidia GeForce 7350LE? Or a 7-series card in general? This is my second computer, and I’m just trying to get a working install of Ubuntu on it. I don’t want to put much money into it, as I have seen Ubuntu run great on much older/less capable machines. I’ve got a decent Intel processor (2.3Ghz), 2GB of RAM, 320GB hard drive, 32-bit architecture, and there is no other O/S installed. It appears as if the graphics card is holding me back. Should I just buy a cheap graphics card (non-NVIDIA) to put in as a replacement? TRIED SO FAR: -all drivers available in Ubuntu *all fail -manual install of some different NVIDIA drivers *all fail -also tried installing the generic kernel, Nvidia driver doesn't work in 12.10 *no difference -every method suggested to at least get a command line after switching to a NVIDIA driver *all fail

    Read the article

  • Protected flash video (requires HAL) on Gentoo

    - by Mala
    I am unable to play "protected" flash video, such as Amazon Prime Instant Video. From what I've read and uncovered, this seems to be due to a lack of HAL being installed on my computer. Confirmation that it is required for protected video can be seen towards the beginning of http://helpx.adobe.com/x-productkb/multi/flash-player-11-problems-playing.html However, hal is not in the gentoo portage tree, and in any case has been deprecated and replaced by udev. How can I go about getting Amazon Prime Instant Video to work again? I was considering grabbing the source from http://www.freedesktop.org/wiki/Software/hal but the links there won't load, and trying to install it from old ebuilds or from overlays which claim to still support it (e.g. kde-sunset) result in a compilation error: In file included from addon-generic-backlight.c:38:0: /usr/include/glib-2.0/glib/gmain.h:21:2: error: #error "Only <glib.h> can be included directly." Has anyone else solved this issue?

    Read the article

  • Random compositing lag

    - by user1020567
    My laptop specs: 512 mb of RAM, out of which 64 mb are shared with an integrated GPU - ATI Radeon Xpress 200 M. Intel 1,6 Ghz Celeron M single-core processor. I've spent months trying to figure out why compositing and effects sometimes lag on any distro I try. Now I've come to realise that no matter what drivers I try (the default ones work for me on pretty much any linux) compositing lag is random. When I used Ubuntu 10.10, for example, sometimes window compositing would lag and sometimes it wouldn't. The PC is able to render those effects so hardware is not the problem. It's completely random and unpredictable - sometimes when I turn on the computer the effects lag horribly and sometimes it's completely smooth. I've also checked startup items and there doesn't seem to be any unnecessary entries. I also tried building my own OS with Arch Linux and the problem persists there, therefore I can only assume that it's a driver issue of some sort. By default there are lots of drivers supplied with linux distributions. Could it be that they're in the way? The ones that I need are ati/radeon (or both? What's the difference between them?) and there seem to be a lot of others... What should I do?

    Read the article

  • I cannot connect to database from Drupal

    - by Patrick
    hi, I've uploaded my drupal website (and related database) to my new server. The database info is: host: localhost user: user pass: pass databaseName = database_name I've set the following line in settings.php file: $db_url = 'mysqli://user:password@localhost/database_name'; but what I get is this: If you are the maintainer of this site, please check your database settings in the settings.php file and ensure that your hosting provider's database server is running. For more help, see the handbook, or contact your hosting provider. I guess the database is running, it always run and I can access with phpmyadmin so I think the problem is not there. The database and website files upload have also been succesfull.. so I dunno what to do to fix this issue. It is mysql on IIS Server thanks

    Read the article

  • Can I use a micro ec2 instance as a load balancer for my other large ec2 instances?

    - by Ryan Detzel
    The issue I'm having is I want to upgrade that instance often(security patches, etc) but I'm affriad something will fail and the site will be down. So, I want to have another server setup and load balance between the two that way I can easily disable one, upgrade it and once it's working add it back in the mix and repeat. What kind of machine is needed for a load balancer? Would the micro instance work just fine? The site gets anywhere from 3-10k hits/day. I plan on using nginx as the load balancer.

    Read the article

  • Is it better to have multiple domains for cities or one single TLD?

    - by Brett
    I make websites for small businesses, and for some reason business owners love to have several domains with the same website but the TLD containing the city name. For example: 1. smallbizname.com 2. clevelandsmallbizname.com 3. columbussmallbizname.com 4. cincinnatismallbizname.com ... and so on. I've seen questions about localization per country aspects, but this is a much smaller scale, so I don't think the same rules apply. The problem I have is the companies never want to write separate content per domain, just have the same website hosted several times at each domain. I feel this probably hurts SEO for two reasons: 1. Traffic gets scattered throughout domains, could be boosting just one domain. 2. Duplicate content penalty because the content is identical. My question boils down to this... Should I redirect all the city domains to the main business name domain, or does having these separate sites help to rank better per city? And if they are redirected, how does google rank the redirects? Thanks for any input on this issue!!

    Read the article

  • Is constantly looking for code examples a sign of a bad developer?

    - by Newly Insecure
    I am a CS student with several years of experience in C and C++, and for the last few years I've been constantly working with Java/Objective C doing app development and now I have switched to web development and am mainly focused on ruby on rails and I came to the realization that (as with app development , really) I reference other code way too much. I constantly Google functionality for lots of things I imagine I should be able to do from scratch and it's really cracked my confidence a bit. Basic fundamentals are not an issue, I hate to use this as an example but I can run through javabat in both java/python at a sprint - obviously not an accomplishment and but what I mean to say is I have a strong base for the fundamentals I think? I know what I need to use typically but reference syntax constantly. Would love some advice and input on this, as it has been holding me back pretty solidly in terms of looking for work in this field even though I'm finishing my degree. My main reason for asking is not really about employment, but more that I don't want to be the only guy at a hackathon not hammering out nonstop code and sitting there with 20 Google/github tabs open, and I have refrained from attending any due to a slight lack of confidence... Is a person a bad developer by constantly looking to code examples for moderate to complex tasks?

    Read the article

  • Snow Leopard Hangs at Login Window

    - by jessecurry
    I've had an issue for the past few months, but I rarely restart so it hasn't caused too much trouble. Basically, when I start up my Mac (iMac10,1 - 3.06GHz Intel Core 2 Duo, OS X 10.6.3) everything proceeds as usual until I reach the login window. The login window displays normally, but keyboard and mouse input seem to be ignored. This condition persists for around 5 minutes at which time everything goes back to normal. While the login window is frozen my second monitor appears entirely blue, the second monitor receives a background as soon as the login window becomes responsive. If I startup while holding SHIFT the problem still occurs, but the freeze is much shorter. Looking through my logs I see no activity during the time that the login window is frozen. I've attempted to repair disk permissions, and gone through every possible maintenance option in Cocktail.

    Read the article

  • What happens when router has been set to incorrect time?

    - by iamrohitbanga
    I have a D-Link router for my home Wi-Fi network. Everyday at least once the internet suddenly goes down. I am simply not able to connect to the Wi-Fi network. If I just restart the router, it starts working. To debug the issue I logged into the admin panel and noticed the time was set to something in 2002. I have set it to the correct time. Will wait to see if that fixes the problem. In the meanwhile I want to know what can go bad when the router has been set to show an incorrect time? What are the kinds of problems expected? My Wi-Fi was working just fine most of the time, but sometimes it lost the connection. Could this be linked to the incorrect time setting?

    Read the article

  • Missing 16:10 resolutions with Nvidia drivers (Can't add resolutions)

    - by Wuinny
    Hello, I have a laptop with a Nvidia 9650M GT and used the drivers that Seven brought me. It works fine but Metro 2033 tells me that i have to upgrade my drivers to play the game. So i did it. But since i did a clean install of the new Nvidia drivers, i just have 1440*900 or 4:3 resolutions. I usually played with 1280*800 or 1184*740 (for performance issue) With the "old" drivers i was able to create custom resolution (1184*740) in Nvidia control panel but now when i try it tells me that "my monitor cannot support this resolution". When i insist, it works but soon as i shut down my computer i have to recreate it.. Do anyone have a fix ? Thank you

    Read the article

  • BES 5.x issues when connecting to Exchange 2010 SP2 RU4

    - by makerofthings7
    Ever since we updated from SP1 RU4 to SP2 RU4 we have noticed that our BES devices will simply stop receiving email. This has occurred at least 5 times in the past few weeks. Today, while speaking with RIM support (a T3 contract), they said that they have reports of sporadic connectivity issues from some BES users. This has been occurring with Exchange 2010 SP2 RU4, and have no public information to give us. We were able to temporarily get things running by doing any or all of the following Restart the BES server Recreate the MAPI profile (deleted the keys in the registry) Restart the CAS servers Wait it out (2 hours) What other options should I take to resolve this issue?

    Read the article

  • The good SQL database to process a lot of data?

    - by Dorian
    I have to process like 10-100 millions records. I have to give the data to the client when it's finish. The data is givent as SQL requests to execute in the database. He have a powerful server with MySQL, I think it will be fast enough. The issue is my computer is not as powerful as his server, so I would like to use an other SQL server who is compatible (I export his database and import it in my computer) with MySQL but more powerful. What should I use? Or am I doomed to use MySQL?

    Read the article

  • 'Buy the app' landing page implementations

    - by benwad
    My site (using Django) has an app that I'm trying to push - I currently have a piece of middleware that redirects the user to a page advertising the app if they're accessing the page on the iPhone, then setting a cookie so that the user isn't bugged by the message every time they visit the site. This works fine, however checking the page with the mobile Googlebot checker shows that the Googlebot gets stuck in the redirect (since it doesn't store cookies) and therefore won't index the proper content. So, I'm trying to think of an alternative implementation that won't hurt the site's Google ranking and won't have any other adverse effects. I've considered a couple of options: Redirect (the current solution), but don't redirect if the user agent matches the Googlebot's UA string. This would be ideal, however I'm not sure if Google like their bot being treated differently from other users, and I'm afraid the site's ranking may be somehow penalised if I go ahead with this. Use a Javascript popup instead of a redirect. This would make sure the Googlebot finds the content it needs, however I envision this approach causing compatibility issues with the myriad mobile devices/browsers out there, and may affect the page load time. How valid are these options? And is there a better option for implementing this feature out there? I've tried researching this topic but surprisingly can't find any reputable-looking blog posts that explore this topic. EDIT: I posted this on SF because it seemed unsuitable for SO, but if there's another site that would be better for this issue then I'd be happy to move the question elsewhere.

    Read the article

  • GillSans font doesn't show up in any applications

    - by roboknight
    I am running Ubuntu 10.04, Lucid Lynx, and I'm trying to add some fonts for use with Libre Office. In particular, I've placed a GillSans.ttc font file into /usr/share/fonts/truetype by copying it there with: sudo cp GillSans.ttc /usr/share/fonts/truetype and then attempting to "install" it by updating the font cache: sudo fc-cache -fv That appears to complete successfully, however, the font never seems to be available to either Libre Office, nor fontmatrix. In addition, I tried adding the font to my .fonts directory, thinking that I was somehow seeing a cache of just user supplied fonts, but that isn't the case. The file is a .ttc file, is there some other tools I need to use to extract each font from the .ttc, or is there something else I must do to use a .ttc? Nothing appears to have any issue with the file, but it could be silently failing to do anything. I've looked for the answer in other places, but the only time that a similar question was asked, it did not appear to be answered, so I figured I would ask the question differently and see if anyone else might be able to clue me in on how to use a .ttc font file, if that is possible. A related question, are there any tools to see what font families/font names are contained within a .ttc font file?

    Read the article

  • Can I ping via an arbitrary interface of a DD-WRT system? [migrated]

    - by bytebuster
    There's a Linksys WRT54GL router with DD-WRT firmware (v23SP2). The network has a simple dual-WAN configuration (standby mode, switching by a script): ~ # ip route 192.168.3.0/24 dev br0 proto kernel scope link src 192.168.3.1 192.168.2.0/24 dev vlan2 proto kernel scope link src 192.168.2.2 192.168.1.0/24 dev vlan1 proto kernel scope link src 192.168.1.67 127.0.0.0/8 dev lo scope link default via 192.168.2.1 dev vlan2 I'm trying to ping a certain server arbitrary via vlan1 or vlan2. What I tried, as suggested here: ping -I vlan2 <address> ping 192.168.2.1 <address> In both cases ping simply exits with no error messages. Also, ping ignores many other parameters, again, by exiting silently. I failed to find any references that DD-WRT has a limited version of ping whatsoever. I also don't think it can be a permissions issue as mentioned here since the only user with DD-WRT is root. What's wrong?

    Read the article

  • Outlook keeps dropping 'Other Calendars'

    - by David Gard
    We run Exchnage 2010 with Outlook 2007, and we have public calendars that users are able to 'Add to favorites...' - these are then shown under 'Other Calendars' when a user clicks their 'Calendar' tab. The issue we are having is that these calendars drop off every now and again. Looking under 'Public Folders - Favorites', I can see that the calendars are there, but Outlook doesn't seem to know that they are calendars any more. Simply clicking on one of them to open it changes the icon to that of a 'Folder containing calendar items' for each of the favorite items, and closing and reopening Outlook gets them back under 'Other Calendars' My question is this - how can I make Outlook stop forgetting that these public folders are calendars? Thanks.

    Read the article

  • Two components offering the same functionality, required by different dependencies

    - by kander
    I'm building an application in PHP, using Zend Framework 1 and Doctrine2 as the ORM layer. All is going well. Now, I happened to notice that both ZF1 and Doctrine2 come with, and rely on, their own caching implementation. I've evaluated both, and while each has its own pro's and cons, neither of them stand out as superior to the other for my simple needs. Both libraries also seem to be written against their respective interfaces, not their implementations. Reasons why I feel this is an issue is that during the bootstrapping of my application, I have to configure two caching drivers - each with its own syntax. A mismatch is easily created this way, and it feels inefficient to set up two connections to the caching backend because of this. I'm trying to determine what the best way forward is, and would welcome any insights you may be able to offer. What I've thought up so far are four options: Do nothing, accept that two classes offering caching functionality are present. Create a Facade class to stick Zend's interface onto Doctrine's caching implementation. Option 2, the other way around - create a Facade to map Doctrine's interface on a Zend Framework backend. Use multiple-interface-inheritance to create one interface to rule them all, and pray that there aren't any overlaps (ie: if both have a "save" method, they'll need to accept params in the same order due to PHP's lack of proper polymorphism). What option is best, or is there a "None of the above" variant that I'm not aware of?

    Read the article

  • File Transfer Protocol (FTP) Service won't install

    - by user64204
    Hi all, Problem description: -I'm running Windows Server 2003 and I'm trying to install the FTP service following this procedure: http://support.microsoft.com/kb/323384 -At step 7 of the above procedure I provide Windows Small Business Server 2003 disc1 and the installation completes successfully (note: the installer looks for 'ftpctrs2.dll' but the disk contains 'FTPCTRS2.DL_' and seems happy with it as per allowed file types). -However the installation silently fails because if I go back to the list of windows components, FTP is still unchecked, and if I open IIS, the FTP sites section is missing. Question: -How can I force the installation of FTP? Additional info: I don't think this is related but will mention it in case I'm wrong and it is: before I could attempt to install FTP I ran into, and fixed, the issue described in this solution: http://support.microsoft.com/kb/838860 Thanks.

    Read the article

  • Installing cURL on Ubuntu

    - by davykiash
    Am trying to install cURL on my ubuntu server using the command sudo apt-get install php5-curl However i get the following error Reading package lists... Done Building dependency tree Reading state information... Done Some packages could not be installed. This may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of Incoming. The following information may help to resolve the situation: The following packages have unmet dependencies: php5-curl: Depends: phpapi-20060613+lfs Depends: php5-common (= 5.2.6.dfsg.1-3ubuntu4.5) but 5.3.2-0.dotdeb.1 is to be installed E: Broken packages I am running PHP Version 5.3.2-0.dotdeb.1 on my server. Whats the issue? I need to get curl up and running.

    Read the article

  • Corrupt Vista Account - With a twist

    - by oldSkool-Soldier
    I have a Windows Vista system whose profile seems to have been corrupted. Many solutions seem to indicate that you create another admin account, log in as that account, and delete the offending profile information for the account that is corrupted. Here's the twist: There are two accounts on this System - 1. Administrator Account 2. Standard Account. It's the Administrator Account which has become corrupted. If you attempted to do any user managment / attempt to create a new account under the 'Standard Account', UAC prompts for a user/pass. It has the admin account come up by default, and asking for a password (of which, there is none). When you click 'OK' - nothing happens (no doubt due to the fact that the account is corrupted, so it can't verify who or what is trying to allow the administrative privilage). Can anyone think of a way of fixing this account, or creating another administrator account to resolve this issue? Many thanks in advance.

    Read the article

  • LAMP server has gone down a few times. Ideas for server optimization?

    - by MattB
    Hi all, Our production web server has gone down a few times over the course of the last half year. In the end, we've needed to contact the web host and have them restart as I'm unable to even SSH in. This appears to only affect the web server and not the MySQL database server which is separate. When it affects the web server, all hosted websites time out. I'd like to examine web server optimization/corrections to get to the root of this issue. Any recommendations on how to proceed with that? I'm sure log files would play a role. I'm able to find my way around a Linux-based server and make needed changes, but would be interested in any tips I may not have thought of yet. It may be best for us to speak with an outside consultant as another option. Thanks.

    Read the article

  • 301 redirect from HTTP to HTTPS - how to be sure Google is fetching the correct information?

    - by user33692
    I'm hoping somebody might be able to provide a bit of advice on an issue I am having. I have one site where we implemented a 301 redirect on the homepage from HTTP to HTTPS. We have links on the homepage to other parts of the site that are not under SSL (in fact there is only one other page under SSL). When I go to our Webmaster Tools account I notice that we are not being provided with any webmaster information (e.g., search queries, backlinks, etc...) related to our homepage under SSL. I performed a Fetch as Google on the homepage and the information it returned is: HTTP/1.1 301 Moved Permanently Date: Fri, 08 Nov 2013 17:26:24 GMT Server: Apache/2.2.16 (Debian) Location: https://mysite.com/ Vary: Accept-Encoding Content-Encoding: gzip Content-Length: 242 Keep-Alive: timeout=15, max=100 Connection: Keep-Alive Content-Type: text/html; charset=iso-8859-1 <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> <html><head> <title>301 Moved Permanently</title> </head><body> <h1>Moved Permanently</h1> <p>The document has moved <a href="https://mysite.com/">here</a>.</p> <hr> <address>Apache/2.2.16 (Debian) Server at mysite.com</address> </body></html> I am worried by the fact that Google fetch is not getting the correct Title tags and Meta information from our homepage and that this is hurting our search results. Additionally, I am worried that we need to do something specific with the sitemap to ensure that Google is correctly indexing all our pages and being able to flow from the HTTPS to the HTTP without issues. Does anybody have any advice on how we can correctly set this up or be sure that Google is fetching the correct information?

    Read the article

  • Using nginx as a reverse proxy for tomcat results in new jsessionids for every ssl request

    - by user439407
    I am using nginx as a reverse proxy for a tomcat setup, and everything works fine for the MOST part, the only issue I am having is that every request to an http address results in a new JSESSION ID being created(this doesn't happen in http), here is the relevant part of the NGINX configuration: location / { proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header Host $http_host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-Proto https; proxy_redirect off; proxy_connect_timeout 240; proxy_send_timeout 240; proxy_read_timeout 240; proxy_pass http://localhost:8080; } Any idea why I am constantly genning new jsessionids?

    Read the article

  • How to use chain.p7b with Apache?

    - by Debianuser
    I wanted to setup a SSL website on Apache and applied for a certificate from my local ISP. All they sent me was a single file named chain.p7b. I have always used certificates from other vendors without any issues but they usually provide two files to be configured as SSLCertificateFile and SSLCertificateChainFile in Apache. Following instructions from several online resources, I opened the p7b file in Windows and extracted 4 certificates from the file. I then tried configuring Apache with one of the files and it worked, but shows a warning: The certificate is not trusted because no issuer chain was provided. I though I have to use remaining 3 files as SSLCertificateChainFile and/or SSLCACertificateFile. I tried that but it didn't work so I am assuming it might be something completely different. Anyone faced this issue before? The following page http://www-01.ibm.com/support/docview.wss?uid=swg21458997 talks about using a keystore but is that relevant to Apache?

    Read the article

  • After deleting a local machines offline file cache, the same user's "my documents" no longer redirects to the network location.

    - by stead1984
    One of my apprentices was tasked with clearing out unused local profiles and clearing the offline file cache. After he cleared the offline file cache and rebooted the machine, he would log in as himself and no longer have his "my documents" redirected to the set network location. More over this seemed to then affect ANY other networked machine he logged into, except his own laptop. All our standard workstations run Windows XP Service Pack 3, the apprentice's laptop runs Windows 7 Professional. I can understand how clearing the offline file cache after deleting old local profiles could cause this issue but draw a complete blank as to why it would affect all networked machines. It's a strange one so this question may be a little hard to understand so any questions or further understanding required please ask.

    Read the article

< Previous Page | 576 577 578 579 580 581 582 583 584 585 586 587  | Next Page >