Search Results

Search found 27606 results on 1105 pages for 'javascript disabled'.

Page 841/1105 | < Previous Page | 837 838 839 840 841 842 843 844 845 846 847 848  | Next Page >

  • Still can't mount windows 8 drive after restart

    - by Ishai Bloch
    After following the instructions in other posts, I am still getting the same error when I try to mount my Windows 8 drive in Ubuntu 14.04 on a dual boot system. I have disabled fast start after shutdown, hybrid hibernation, and the preinstalled Asus Instant On service. I have tried restarting Windows rather than shutting down. In all cases I get the same error message, namely: Error mounting /dev/sda4 at /media/jesse/OS: Command-line `mount -t "ntfs" -o "uhelper=udisks2,nodev,nosuid,uid=1000,gid=1000,dmask=0077,fmask=0177" "/dev/sda4" "/media/jesse/OS"' exited with non-zero exit status 14: Windows is hibernated, refused to mount. Failed to mount '/dev/sda4': Operation not permitted The NTFS partition is in an unsafe state. Please resume and shutdown Windows fully (no hibernation or fast restarting), or mount the volume read-only with the 'ro' mount option. I did not have the same issue before upgrading from Ubuntu 12 to 14. For what it's worth my computer is supposedly a "hybrid" with an SSD drive installed, although I can't see that the SSD drive is being used at all with my present settings. Any thoughts?

    Read the article

  • My internet connection slows or dies unexpectedly

    - by genesis
    I installed Ubuntu 10.04 once again and I'm having some problems which I had before, but I have no idea how I solved them. On Windows, everything's working fine and I had no problems with this. My problem is that sometimes, when browsing through the internet, webpages just start to load really slow, sometimes it doesn't load anything at all (Error 118 (net::ERR_CONNECTION_TIMED_OUT): The operation timed out.) and it starts to work after few minutes. My IPv4 settings are automatic (DHCP), and IPv6 settings are Ignored/Disabled. I think my previous problems had something to do with IPv6, but I'm not sure. Is there a fix for this? iwconfig lo no wireless extensions. eth0 no wireless extensions. wlan0 IEEE 802.11bgn ESSID:"Fsite1" Mode:Managed Frequency:2.442 GHz Access Point: C8:3A:35:40:43:68 Bit Rate=0 kb/s Tx-Power=20 dBm Retry long limit:7 RTS thr:off Fragment thr:off Encryption key:off Power Management:on Link Quality=43/70 Signal level=-67 dBm Rx invalid nwid:0 Rx invalid crypt:0 Rx invalid frag:0 Tx excessive retries:0 Invalid misc:0 Missed beacon:0

    Read the article

  • Electricsheep Not Working in Raring 13.04

    - by Nhoeh37
    I just downloaded Electricsheep four days before I updated my Ubuntu software and I fell in love with it. After I upgraded to 13.04, I'm kind of upset in saying that Electricsheep no longer works. I have no idea how to fix it and I feel helpless with Linux. As I said, it was working completely fine in 12.10, but now with the Raring Ringtail update: I use Xscreensaver to launch Electricsheep. When I open Xscreensaver, "Electricsheep" is one of the options to choose. When I "preview" ES or even launch the "xscreensaver-demo" in the terminal, it doesn't work anymore; I get an error message. The error message says: "detected another electric sheep process. rendering of sheep is disabled using read only access." Damnit, can someone please help me? I want my trippy screensaver back, it shouldn't be this hard! EDIT: I contacted the creator of Electricsheep (Scott Draves) via the link to his email address on his personal website and asked for his assistance. Hopefully the Askubuntu community will be able to fix this problem, I'm sure I am not the only one who is having this problem as I have searched and searched.

    Read the article

  • Cannot use second display with 12.04 and Intel 2000/3000

    - by Carolyn Marenger
    I am unable to get anything to display on my second monitor, or even get the system to recognize that there is a second screen. I am running Ubuntu 12.04 on a Gigabyte GA-H61M-S2PV, revision 2.0 box. The integrated chipset is an Intel 2000/3000, and there are both a D-Dub and a DVI-D display ports on the MB. This is the first operating system I have installed on this system. I have a second monitor plugged into the DVI-D port via a DVI-D to D-Sub adapter. I cannot verify that the motherboard or adapter were/are working, short of installing windows to test the theory. When I go into the "System Settings - Displays" control window, it shows one display. I have rebooted with the second monitor attached, and I have perused the BIOS settings in case it might have been disabled. So far, I have had no indication that the second monitor is recognized, not even a flicker at power on. If I swap monitors and cables between the DVI-D and D-Dub ports, the other monitor lights up, so I know the monitor and video cable are not the issue. Any suggestions would be appreciated. Thanks, Carolyn

    Read the article

  • Maximum 5 minute battery life with Ubuntu 11.10 on HP laptop

    - by JamesG
    I apologise if this question is too similar to the numerous others already asked, but it seems that my difference in battery life is significantly more noticeable than others that have been reported. I recently installed Ubuntu 11.10 on my HP Pavilion dv6 laptop (which I purchased brand new just under one year ago). When running Windows 7 on this laptop, I have been able to get up to two and a half to three hours of battery life with wireless disabled and when running only Microsoft Word. However, when running Ubuntu, I am unable to use the laptop if it is not plugged in. Upon unplugging the fully-charged machine from the power cord, if I have wireless enabled, I immediately receive a notification that the battery levels are critically low and that shutdown is imminent. Even if I replace the power plug, the laptop shuts down within thirty seconds. If I disable wireless capability, I am able to run the laptop for an absolute maximum of five minutes on battery powers before receiving the same message. I have tried running with Jupiter on Power Saver mode, but to no noticeable effect. Ignoring the fact that I can't use my laptop without being attached to a power source, I really do enjoy using Ubuntu, and hence would greatly appreciate any help that can be offered.

    Read the article

  • Compiling custom kernel 3.7.x lowlatency on Ubuntu 12.04

    - by FlabbergastedPickle
    All, I have a peculiar problem with trying to compile a lowlatency flavor of the latest 3.7 kernel. I retrieved the prepatched source from the launchpad using bzr, compiled it using the usual make-kpkg using the current config file plus default options for the rest, installed the kernel and booted into it. Everything works except for the fglrx and wl drivers that I had to install in the original 12.04 lowlatency kernel. So, I tried recompiling these and succeeded with both of them (no errors were reported)--wl driver required a minor adjustment to system.h include while latest fglrx 12.11 beta11 (released yesterday, Dec. 3rd, 2012) compiled without the hitch. Yet, when I try to modprobe either module (both having in common the fact that they were built after the kernel, fglrx as a deb, and wl via the usual make/make install), I get "FATAL: no MODULENAME module found" (MODULENAME being either wl or fglrx). The graphic driver watermark shows 3D crossed out and "for testing purposes" (or "unsupported hardware," can't remember), and no fglrx or wl is loaded. More mysteriously, dmesg shows no attempt on kernel's behalf to load the said drivers, even though they are clearly in the right /lib/modules/KERNEL_VERSION folder. How is this possible? Has something fundamentally changed in 3.7 kernel that would prevent modprobing of these? I know that there is driver signing option that was merged recently but as far as I could tell the kernel config file generated by the build process had that disabled. OTOH, while building wl driver, I did get a warning that the driver was not signed... Then again, even if the kernel disallowed loading of those modules, shouldn't dmesg reflect that? Any thoughts on this one are most appreciated.

    Read the article

  • Enterprise Instrumentation: The 'sessionName' parameter of value 'TraceSession' is not valid

    - by Michael Freidgeim
    We are still using Enterprise Instrumentation(that was created during .Net 1.1 time)In new Server 2008 environment and IIS 7 we have the following errors:The 'sessionName' parameter of value 'TraceSession' is not valid. A trace session of this name does not exist in the TraceSessions configuration file for Windows Trace Session Manager service. Ensure that a session of this name exists in the TraceSessions configuration file and that the Windows Trace Session Manager service is started.   at Microsoft.EnterpriseInstrumentation.EventSinks.TraceEventSink..ctor(IDictionary parameters, EventSource eventSource)   --- End of inner exception stack trace ---   at System.RuntimeMethodHandle._InvokeConstructor(IRuntimeMethodInfo method, Object[] args, SignatureStruct& signature, RuntimeType declaringType)   at System.Reflection.RuntimeConstructorInfo.Invoke(BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)   at System.RuntimeType.CreateInstanceImpl(BindingFlags bindingAttr, Binder binder, Object[] args, CultureInfo culture, Object[] activationAttributes)   at Microsoft.EnterpriseInstrumentation.EventSinks.EventSink.CreateNewEventSinks(DataRow[] eventSinkRows, EventSource eventSource)I’ve seen the same errors on development Win7 machines when using IIS. It seems not a problem on Cassini.I've checked ,that Windows Trace Session Manager Service has started and The file C:\Program Files (x86)\Microsoft Enterprise Instrumentation\Bin\Trace Service\TraceSessions.config has corresponding entry<?xml version="1.0" encoding="utf-8" ?><configuration >                <defaultParameters minBuffers="4" maxFileSize="10" maxBuffers="25" bufferSize="20" logFileMode="sequential" flushTimer="3" />                <sessionList>                                 <session name="TraceSession" enabled="false" fileName="C:\Program Files (x86)\Microsoft Enterprise Instrumentation\Bin\Trace Service\Logs\TraceLog.log" />                </sessionList></configuration>The errors still continue, but I was able to disable  the parameter in  eventSink configuration   <eventSink name=" traceSink" description=" Outputs events to the Windows Event Trace." type ="Microsoft.EnterpriseInstrumentation.EventSinks.TraceEventSink ">                <!-- MNF disabled parameter to  avoid error "The 'sessionName' parameter of value 'TraceSession' is not valid"                      < parameter name ="sessionName " value ="TraceSession " />                    -->    </ eventSink>Related old post http://bytes.com/topic/net/answers/104761-enterprise-instrumentation-windows-trace-session-managerOne day I wish to replace all EnterpriseInstrumentation calls with NLog.

    Read the article

  • How to efficiently protect part of an application with a license

    - by Patrick
    I am working on an application that has many functional parts. When a customer buys the application, he buys the standard functionality, but he can also buy some additional elements of the application for an additional price. All of the elements are part of the same application executable. A license key is used to indicate which of the elements should be accessible in the application. Some of the elements can be easily disabled if the user didn't pay for it. These are typically the modules that you can access via the application's menu. However, some elements give more problems: What if a part of the data model is related to an optional part? Do I build up these data structures in my application so the rest of my application can just assume they're always there? Or do I don't build them, and add checks in the rest of may application? What if some optional part is still useful to perform some internal tasks, but I don't want to expose it to the user externally? What if the marketing responsible wants to make a standard part now an optional part? In all of my application I assume that that part is present, but if it becomes optional, I should add checks on it everywhere in the application. I have some ideas on how to solve some of the problems (e.g. interfaces with dual implementations: one working implementation, and one that is activated if the optional part is not activated). Do you know of any patterns that can be used to solve this kind of problem? Or do you have any suggestions on how to handle this licensing problem? Thanks.

    Read the article

  • Xorg becomes unkillable at 3AM

    - by chew socks
    Most nights, some time in the hour of 3AM my xorg process will increase to 100% cpu and gpu load will also increase to 100%. The process also becomes unkillable. I cannot sudo kill -9 it or get back control with sudo service lightdm restart. I also cannot switch to to a tty screen with ctrl + alt + f1. To reboot I have to log in with ssh, but this is not perfect because if I reboot while it is doing this my ZFS pool will fail to mount when it comes back up ( that is where my /home is ). Does anyone have any ideas as to why I can't stop and restart xorg, or even better, know why this is happening? Thanks NOTE: For anyone who comes looking for the same problem. I disabled catalyst AI and made it through the night. I've been up for 1 day 3 hours now. My record for this month is 2 days and 19 hours without a problem. My all time record is 6 days without a crash. I'll post here if it crashes again or I'm able to set a new record.

    Read the article

  • How can I get nvidia-96 installed?

    - by Bob
    I'm at my wits end here. This is my last effort before I go back to Windows. I need to get nvidia-96 proprietary driver installed. Synaptic won't install it because it says it has dependencies. I installed every single dependency it listed except for "xorg-video-abi-10" which does not show up as an item that can be installed. I have no idea what to do. Using 11.10 with a NVIDIA Geforce 3 GPU. Anyone know how to get this dang driver installed? @fossfreedom: the opensource driver is extremely slow. So slow that the OS is unusable—words appear seconds after I type them—programs take forever to perform actions. Also it is causing my monitor to turn on and off for no reason. @yossile: synaptic shows that I have xserver-xorg-core installed. And xserver-xorg-core-udeb does not show up as something that can be installed. @papseddy: when I try to install the downloaded nvidia driver it says it won't work until I disable Nouveau kernel driver. I have tried everything to get this dang Nouveau kernel driver disabled. Nothing has been successful.

    Read the article

  • Updated Virtual Machine for VS/TFS 2010

    - by Enrique Lima
    If you had downloaded the previous version of the virtual machines, then you are likely aware they are set to expire soon (12/15/2010). Brian Keller announced yesterday (blog post here) the availability of a vm refresh (new expiration set for 6/1/2011). What is part of the refresh? Here is the excerpt from Brian’s post: “ The version of this virtual machine which was refreshed on December 9, 2010, includes the following additions: · Visual Studio 2010 Feature Pack 2 · Team Foundation Server 2010 Power Tools (September 2010 Release) · Visual Studio 2010 Productivity Power Tools (these are disabled in VS so that the screenshots of the hands-on-labs still match; you can quickly enable the Productivity Power Tools via Tools -> Extension Manager from within Visual Studio) · Test Scribe for Microsoft Test Manager · Visual Studio Scrum 1.0 Process Template · All Windows Updates through December 8, 2010 · Lab Management GDR (KB983578) · Visual Studio 2010 Feature Pack 2 pre-requisite hotfix (KB2403277) · Microsoft Test Manager hotfix (KB2387011) · Minor fit-and-finish fixes based on customer feedback · A new expiration date of June 1, 2011” The links to download the Virtual Machines are: Hyper-V: http://www.microsoft.com/downloads/en/details.aspx?FamilyID=e0198b64-4acb-4709-b07f-359fb4d523bc&displaylang=en Windows Virtual PC (Win 7): http://www.microsoft.com/downloads/en/details.aspx?FamilyID=509c3ba1-4efc-42b5-b6d8-0232b2cbb26e&displaylang=en

    Read the article

  • Notification framework for object lifecycle

    - by rlandster
    I am looking for an application, framework, or library that would help us with "object life-cycle management". There are many things that are created for users, departments, and services that, all too often, are left unmanaged. Some examples: user accounts groups SSL certificates access rights databases software license provisionings storage list-serve accounts These objects are created and managed by a wide variety of applications and systems. Typically, a user (person) requests (either explicitly or implicitly) one of these objects. A centralized management tool would help us manage such administration chores as: What objects does user X currently own/manage? Move the ownership of object P to user X; move all objects owned by user X (who was just been fired) to user Y. For all objects of type T that have expired be sure the objects have been disabled or deleted by their provider. How many active (expired, about-to-expire) objects of type P are there? Send periodic notifications to all users who own active objects of type P reminding them of what they own. There is a security alert for objects of type P; send a notification to all users who own these types of objects to take a specific remedial action. Delete or disable a set of objects based on expiration (or some other criteria). These objects are directly managed through their own applications (Active Directory, MySql, file systems, etc.) and may even have their own notification systems, but I want to centralize this into an "object management system". The OMS should allow the association with an external identity provider that defines who the users and groups are (e.g., LDAP, Active Directory) creation of objects association of an object to a specific user and/or group association with an expiration date creation of flexible reporting including letting users know what objects they currently own and their expiration dates integration with an external object "provider" via a plug-in We could write something from scratch, but I am hoping there is something already out there that will help, either an entire application or a set of libraries that provide much of what is needed. Any ideas?

    Read the article

  • How to tune touchpad for smaller area?

    - by M.L.
    I just purchased the Asus Zenbook UX31A and I require some advice regarding tuning the touchpad. Basically my problem is that the mouse buttons are withing the touch area and I would like it to be disabled. Such that only part of the touch area is recognized. I have attached an image to clarify the idea. Basically only the red area in the image should be recognized. I have two reasons for this request. Firstly, I like to hold my thumb on the mouse button instead of holding it in the air. This is problematic, because touchpad thinks that I'm trying some kind of two finger gesture all the time and cursor stops moving. Clicking is also problematic since every finger movement -- regardless of how insignificant it might be -- on the "button" will move the cursor. I have been playing with some synclient settings like AreaBottomEdge etc. They have some effect but then I cannot highlight text or right button stops working. How could I solve this?

    Read the article

  • Dual Monitor 'How To' for 12.04

    - by Kim Prince
    I recently built my own PC and was delighted with the result, except for a problem with dual monitors. After having tried a few different combinations of hardware, I think what I really need is a 'how to' explanation. My motherboard is an MSI Z77MA-G45, which has an analogue, a DVI, and a HDMI port. Initially I hooked monitors up to the DVI and analogue port and it seemed to work fine. Both screens worked independently of each other, it was great. After a few days I started turning my PC off at night, and when I tried to turn it back on it would boot into the terminal mode. I would have to turn one of the monitors off and after rebooting a few times, it would eventually boot into an X Window session. Occasionally I would see an error relating to Xorg. I upgraded the motherboard BIOS but that made no difference. Eventually, I installed a graphics card - an NVIDIA GeForce GT 520. Now it seems that my on board graphics have been disabled completely, and I am reliant on the graphics card. Furthermore, the graphics card seems to only recognise one screen at a time. (The first time I rebooted with both plugged in, it flashed up a message saying that it was auto-selecting DVI). Anyhow, I think I need some 'how to' (or perhaps 'where to'), from here. For example, is X Windows configuration the next place to look? And how do I go about configuring X Windows? (Note that in Systems Settings it says my graphics driver is 'unknown', and when I ask it to detect monitors, it sees only the one!)

    Read the article

  • wireless blocked after installing ubuntu 12.04

    - by Cornelia Frank
    I am using a lenovo S10-3 ideapad; had no problems with earlier version of ubuntu, only since installing 12.04. Have looked through many of the questions on the same issue and tried potential solutions but cannot seem to solve my problem. The hardware switch is in 'on' position and the wireless light comes on very briefly (2-3 sec) when the laptop starts up but then goes off and stays off. Pressing FN+F5 does nothing at all. I'd be grateful for any assistance. Cornelia Have received the following responses in Terminal: cf@cf-Lenovo:~$ rfkill list all 0: ideapad_wlan: Wireless LAN Soft blocked: no Hard blocked: no 1: ideapad_bluetooth: Bluetooth Soft blocked: no Hard blocked: no 2: phy0: Wireless LAN Soft blocked: no Hard blocked: yes cf@cf-Lenovo:~$ iwconfig lo no wireless extensions. wlan0 IEEE 802.11bgn ESSID:off/any Mode:Managed Access Point: Not-Associated Tx-Power=off Retry long limit:7 RTS thr:off Fragment thr:off Power Management:off eth0 no wireless extensions. cf@cf-Lenovo:~$ lshw -C network WARNING: you should run this program as super-user. *-network description: Ethernet interface product: RTL8101E/RTL8102E PCI Express Fast Ethernet controller vendor: Realtek Semiconductor Co., Ltd. physical id: 0 bus info: pci@0000:05:00.0 logical name: eth0 version: 02 serial: 00:26:9e:ee:7f:4c size: 100Mbit/s capacity: 100Mbit/s width: 64 bits clock: 33MHz capabilities: bus_master cap_list rom ethernet physical tp mii 10bt 10bt-fd 100bt 100bt-fd autonegotiation configuration: autonegotiation=on broadcast=yes driver=r8169 driverversion=2.3LK-NAPI duplex=full firmware=N/A ip=10.0.1.8 latency=0 multicast=yes port=MII speed=100Mbit/s resources: irq:43 ioport:2000(size=256) memory:f0520000-f0520fff memory:f0510000-f051ffff memory:f0540000-f055ffff *-network DISABLED description: Wireless interface product: AR9285 Wireless Network Adapter (PCI-Express) vendor: Atheros Communications Inc. physical id: 0 bus info: pci@0000:09:00.0 logical name: wlan0 version: 01 serial: c4:17:fe:f8:bc:d7 width: 64 bits clock: 33MHz capabilities: bus_master cap_list ethernet physical wireless configuration: broadcast=yes driver=ath9k driverversion=3.2.0-31-generic-pae firmware=N/A latency=0 multicast=yes wireless=IEEE 802.11bgn resources: irq:18 memory:f0100000-f010ffff WARNING: output may be incomplete or inaccurate, you should run this program as super-user.

    Read the article

  • Using HTML5 Today part 3&ndash; Using Polyfills

    - by Steve Albers
    Shims helps when adding semantic tags to older IE browsers, but there is a huge range of other new HTML5 features that having varying support on browsers.  Polyfills are JavaScript code and/or browser plug-ins that can provide older or less featured browsers with API support.  The best polyfills will detect the whether the current browser has native support, and only adds the functionality if necessary.  The Douglas Crockford JSON2.js library is an example of this approach: if the browser already supports the JSON object, nothing changes.  If JSON is not available, the library adds a JSON property in the global object. This approach provides some big benefits: It lets you add great new HTML5 features to your web sites sooner. It lets the developer focus on writing to the up-and-coming standard rather than proprietary APIs. Where most one-off legacy code fixes tends to break down over time, well done polyfills will stop executing over time (as customer browsers natively support the feature) meaning polyfill code may not need to be tested against new browsers since they will execute the native methods instead. Your should also remember that Polyfills represent an entirely separate code path (and sometimes plug-in) that requires testing for support.  Also Polyfills tend to run on older browsers, which often have slower JavaScript performance.  As a result you might find that performance on older browsers is not comparable. When looking for Polyfills you can start by checking the Modernizr GitHub wiki or the HTML5 Please site. For an example of a polyfill consider a page that writes a few geometric shapes on a <canvas> <script src="jquery-1.7.1.min.js"><script> <script> $(document).ready(function () { drawCanvas(); }); function drawCanvas() { var context = $("canvas")[0].getContext('2d'); //background context.fillStyle = "#8B0000"; context.fillRect(5, 5, 300, 100); // emptybox context.strokeStyle = "#B0C4DE"; context.lineWidth = 4; context.strokeRect(20, 15, 80, 80); // circle context.arc(160, 55, 40, 0, Math.PI * 2, false); context.fillStyle = "#4B0082"; context.fill(); </script>   The result is a simple static canvas with a box & a circle:   …to enable this functionality on a pre-canvas browser we can find a polyfill.  A check on html5please.com references  FlashCanvas.  Pull down the zip and extract the files (flashcanvas.js, flash10canvas.swf, etc) to a directory on your site.  Then based on the documentation you need to add a single line to your original HTML file: <!--[if lt IE 9]><script src="flashcanvas.js"></script><![endif]—> …and you have canvas functionality!  The IE conditional comments ensure that the library is only loaded in browsers where it is useful, improving page load & processing time. Like all Polyfills, you should test to verify the functionality matches your expectations across browsers you need to support.  For instance the Flash Canvas home page advertises 70% support of HTML5 Canvas spec tests.

    Read the article

  • Assigning a script to a keystroke to toggle touchpad

    - by sodiumnitrate
    Since my default sony vaio shortcuts don't completely work in Ubuntu 12.04, I'd like to assign a script to Fn + F1, which toggles the touchpad on and off, so that the cursor would stop moving while I'm typing. Since I use a mouse and rarely need to use the touchpad, I don't want to use "disable touchpad while writing", which doesn't really seem to work anyway. I figured that using a script with the following command (this works, but I have to open up a terminal each time): xinput set-prop 12 "Device Enabled" 0 I have two problems at this point. One is that I don't know how to write this script so that it will toggle it off if it is on, and on if it is off. I know I should use an if statement but I don't know what value I should be checking to see if it is on or off. The second one is that I am having problems creating a new shortcut. I use System Settings - Keyboard - Shortcuts. I tried to add, to custom shortcuts, a new one by clicking the '+' sign. I named it Toggle Touchpad, and added the path to the executable script with the line above, by typing /home/irem/.toggletouchpad I have made it an executable with chmod. The problem is that when I click apply, and then click back on it to define the keystroke, it re-opens the dialogue. I cannot define new keys. (It says disabled on the right column of the entry). I have also tried xbindkeys, which almost constantly crashes. I'd prefer the system settings, if I can set the shortcut. I'd appreciate if anyone can help. Thanks.

    Read the article

  • kismet on BCM43227

    - by Uttam Baroi
    I am trying to monitor wireless on Broadcom BCM43227, I used sudo airmon-ng to run the monitoring, i get command not found. I installed kismet, when i run, i get this *uttam@UT:~$ sudo kismet Launching kismet_server: //usr/bin/kismet_server Suid priv-dropping disabled. This may not be secure. No specific sources given to be enabled, all will be enabled. Non-RFMon VAPs will be destroyed on multi-vap interfaces (ie, madwifi-ng) Enabling channel hopping. Enabling channel splitting. NOTICE: Disabling channel hopping, no enabled sources are able to change channel. Source 0 (addme): Opening none source interface none... FATAL: Please configure at least one packet source. Kismet will not function if no packet sources are defined in kismet.conf or on the command line. Please read the README for more information about configuring Kismet. Kismet exiting. Done. uttam@UT:~$* I did check a blog about kismet on Broadcom that says about some binary drivers not allowing to do it... I used iwconfig and it says no extension : what is that well I need to give a hand on air monitoring............ help, how to do it

    Read the article

  • I just recursively chmod'd everything under / to 750. Any tips?

    - by Ouairz
    I won't be the first and I won't be the last, I suppose. While playing around with the find command, I made a whoops and it would appear that instead of changing the permissions of the ~/web directory to 750, it changed the permissions of the entire filesystem (/) to 750, however I'm not certain, but any attempt to investigate is thwarted by Permission denied messages. For everything. This was the offending command: sudo find ~/web . type d -exec chmod 750 {} If I'm not mistaken, the Ubuntu team disabled root logins as a safety precaution so I'm out of ideas. I'm (obviously) a total newbie when it comes to file permissions so I was wondering if anyone had some good or even some bad advice to share. I've mentally prepped myself to losing everything on the computer which is only of mild consequence, since I have backups, but I did do a bit of work on this box over the week and it would be a shame to lose it all due to a boneheaded mistake. If you are reading this message, ask yourself, have you backed up any of your work recently? Thanks in advance for any insights. Feel free to scold me for using sudo carelessly

    Read the article

  • Disabling monitor reconfiguration when closing lid

    - by Tomas
    I often need to move my laptop from one working place to another. When I do this, there are two events Ubuntu responds to by changing the monitor set up: Removing/attaching the VGA cable Closing/opening the lid of the laptop While removing the VGA cable gives me what I need (single screen, highest native resolution on the external screen if connected; otherwise highest resolution on the laptop), the laptop close/open lid response is not as good. Every time I close or open the lid, Ubuntu reconfigures the monitor set up. When I close the lid now... the screen goes black for a few seconds and it switches to clone, with my laptop screen disabled. Reopening results in... briefly a black screen, then the external monitor being used as desktop extension. Ubuntu thinks too much. My first and foremost question: Is there any way to let Ubuntu ignore lid close events? Ideally (or when there's no way to solve above question) I'd want to change how it deals with the screen reconfiguration. Why does Ubuntu toggle the screen configuration between external, clone and single display? Can't I just configure it to always use the external monitor, when present, in single screen mode? Note that similar questions have been asked before (most notably this one), but these have been closed perhaps wrongly. Any ideas are very welcome, I don't mind playing around a bit to see if something works.

    Read the article

  • What is the program "Additional Drivers" (jockey-gtk) talking me about?

    - by Robert Vila
    The program says: "No proprietary drivers are in use on this system" But it doesn't say if it is talking about graphical drivers only or what. Then, it lists two drivers: NVIDIA accelerated graphics driver (version 173). NVIDIA accelerated graphics driver (version current) [recomended] Both have exactly the same description. What is the difference then? When I select the 1st one, it says: "This river is not activated",and there's a button to "activate" it. When I select the 2nd one, it says: "This river is activated but it is not currently in use", and the button is to "remove". So which one is in use? Why or what for should I have activated (enabled) and not in use? If it is in use it is activated? What is the difference between activate and remove? and what is the relationship between installed, activated, in use, enabled and removed, disabled, inactive and not-installed? Why can I activate the inactivated and remove (but not deactivate) the activated that is not in use? All this is very puzzling... What other drivers can I use for an Apple MacBook pro 3,1 and how? I see that there's a nouveau and I heard that there was going to be a new open source even better. > -display > description: VGA compatible controller > product: G84 [GeForce 8600M GT] > vendor: nVidia Corporation

    Read the article

  • Two views of Federation: inside out, and outside in

    - by Darin Pendergraft
    IDM customers that I speak to have spent a lot of time thinking about enterprise SSO - asking your employees to log in to multiple systems, each with distinct hard to guess (translation: hard to remember) passwords that fit the corporate security policy for length and complexity is a strategy that is just begging for a lot of help-desk password reset calls. So forward thinking organizations have implemented SSO for as many systems as possible. With the mix of Enterprise Apps moving to the cloud, it makes sense to continue this SSO strategy by Federating with those cloud apps and services.  Organizations maintain control, since employee access to the externally hosted apps is provided via the enterprise account.  If the employee leaves, their access to the cloud app is terminated when their enterprise account is disabled.  The employees don't have to remember another username and password - so life is good. From the outside in - I am excited about the increasing use of Social Sign-on - or BYOI (Bring your own Identity).  The convenience of single-sign on is extended to customers/users/prospects when organizations enable access to business services using a social ID.  The last thing I want when visiting a website or blog is to create another account.  So using my Google or Twitter ID is a very nice quick way to get access without having to go through a registration process that creates another username/password that I have to try to remember. The convenience of not having to maintain multiple passwords is obvious, whether you are an employee or customer - and the security benefit of not having lots of passwords to lose or forget is there as well. Are enterprises allowing employees to use their personal (social) IDs for enterprise apps?  Not yet, but we are moving in the right direction, and we will get there some day.

    Read the article

  • How do I connect to a wireless network?

    - by Keith Groben
    I just installed 10.10 x64 and cannot even find my wireless network let alone connect to it. I've searched all over SE and Ubuntu forums and cannot find out how to do this simple thing. Can some one please give me the answer? It is plugged in right now and is 100% updated. It is a Desktop with wireless card. 0: phy0: Wireless LAN Soft blocked: no Hard blocked: no Here's the output: *-network DISABLED description: Wireless interface product: RT2860 vendor: RaLink physical id: 0 bus info: pci@0000:03:00.0 logical name: wlan0 version: 00 serial: 70:1a:04:f4:de:e9 width: 32 bits clock: 33MHz capabilities: pm msi pciexpress bus_master cap_list ethernet physical wireless configuration: broadcast=yes driver=rt2800pci driverversion=2.6.35-27-generic firmware=N/A latency=0 link=yes multicast=yes wireless=IEEE 802.11bgn resources: irq:17 memory:fcff0000-fcffffff *-network description: Ethernet interface product: RTL8111/8168B PCI Express Gigabit Ethernet controller vendor: Realtek Semiconductor Co., Ltd. physical id: 0 bus info: pci@0000:02:00.0 logical name: eth0 version: 03 serial: 00:23:54:fd:c2:32 size: 100MB/s capacity: 1GB/s width: 64 bits clock: 33MHz capabilities: pm msi pciexpress msix vpd bus_master cap_list rom ethernet physical tp mii 10bt 10bt-fd 100bt 100bt-fd 1000bt 1000bt-fd autonegotiation configuration: autonegotiation=on broadcast=yes driver=r8169 driverversion=2.3LK-NAPI duplex=full ip=192.168.1.14 latency=0 link=yes multicast=yes port=MII speed=100MB/s resources: irq:44 ioport:d800(size=256) memory:fceff000-fcefffff memory:ddffc000-ddffffff memory:fcec0000-fcedffff === Update === I have discovered that this is a know issue with the rt2860. I have been following step by step the instructions found here: http://www.ctbarker.info/2010/05/ubuntu-1004-wireless-chipsets-and-wpa.html I decided to stat over because I was getting stuck on step 5: 'sudo rmmod rt2860sta' is was giving me this problem: 'ERROR: Module rt2860sta does not exist in /proc/modules' Since I started over I cannot even get past step 5 'sudo make' I get this: 'make: * No targets specified and no makefile found. Stop.' I am lost. Any help would be appreciated.

    Read the article

  • Fault Handling Slides and Q&A by Vennester

    - by JuergenKress
    Fault Handling It is one thing to architect, design, and code the “happy flow” of your automated business processes and services. It is another thing to deal with situations you do not want or expect to occur in your processes and services. This session dives into fault handling in Oracle Service Bus 11g and Oracle SOA Suite 11g, based on an order-to-cash business process. Faults can be divided into business faults, technical faults, programming errors, and faulty user input. Each type of fault needs a different approach to prevent them from occurring or to deal with them. For example, apply User Experience (UX) techniques to improve the quality of your application so that faulty user input can be prevented as much as possible. This session shows and demos what patterns and techniques can be used in Oracle Service Bus and Oracle SOA Suite to prevent and handle technical faults as well as business faults. Q&A This section lists answers to the questions that were raised during the preview event. Q: Where can retries be configured in Oracle Service Bus? The retry mechanism is used to prevent faults caused by temporary glitches such as short network interruptions. A faulted message is resend (retried) and might succeed this time since the glitch has passed. Retries are an out-of-the-box feature that can be used in Oracle Service Bus and Oracle SOA Suite using the Fault Policy framework. By default, retries are disabled in Oracle Service Bus. Read the full article. SOA & BPM Partner Community For regular information on Oracle SOA Suite become a member in the SOA & BPM Partner Community for registration please visit  www.oracle.com/goto/emea/soa (OPN account required) If you need support with your account please contact the Oracle Partner Business Center. Blog Twitter LinkedIn Mix Forum Technorati Tags: fault handling,vennester,SOA Community,Oracle SOA,Oracle BPM,Community,OPN,Jürgen Kress

    Read the article

  • Brightness going up to 100% on loading certain websites in Chrome

    - by picheto
    I'm using Google Chrome version 21.0.1180.89 on Ubuntu 12.04 and my laptop is a Sony VAIO VPCCW15FL (spec sheet). My video driver is the propietary "NVIDIA accelerated graphics driver (post-release updates)(version-current updates)". After installing Ubuntu, I discovered that neither the brightness control buttons (hardware) or the brightness slider (software) worked, and found out I could get the hardware buttons to work by installing the nvidiabl.deb package and oBacklight script. I'm using nvidiabl-dkms 0.77 and oBacklight 0.3.8. Still, the slider on the Ubuntu "Settings" does not work, but I don't care. There is an annoying thing happening when loading certain pages in Google Chrome: the brightness goes up to 100% when loading the webpage or when leaving it (closing the tab or typing a different URL on the omnibox). However, the "brightness tooltip" (that default brightness notification) remembers the position it was set to, so if I adjust the brightness with the HW buttons, the level gets adjusted relative to the value it was set to before "going 100%". I disabled the flash PPAPI plugin, but left the NPAPI plugin enabled, and the problem went away for pages with flash content. Still, the same thing happens when viewing HTML5 video, or when loading, for example, the Chrome Web Store or using the Scratchpad extension. I suppose it has to do with the rendering of certain elements using the GPU, but this is just a guess. This brightness thing does not happen when using Firefox 15.0 or any other application I have used yet. Does anybody know why this may be happening and what could I do to fix this without changing browser? Thanks a lot.

    Read the article

< Previous Page | 837 838 839 840 841 842 843 844 845 846 847 848  | Next Page >