Search Results

Search found 24253 results on 971 pages for 'multiple monitor'.

Page 350/971 | < Previous Page | 346 347 348 349 350 351 352 353 354 355 356 357  | Next Page >

  • ERROR_PROCEDURE Does Not Return a Schema Name

    "A recent blog entry I read reminded me again that I wanted to rant about an issue in SQL Server for quite some time now. SQL Server 2005 introduced the separation between user and schema. Though schemata already existed before SQL Server 2005, they really became usable with this version, imho. At the same time TRY...CATCH was a new way for structured error handling introduced. And so it finally became possible…" NEW! SQL Monitor 2.0Monitor SQL Server Central's servers withRed Gate's new SQL Monitor.No installation required. Find out more.

    Read the article

  • How do I change local display resolution via SSH?

    - by krumble1
    I have a machine running Ubuntu Server 12.04 LTS and I'm having problems with the local monitor. The Dell flat panel I'm using only displays this message while the computer running: "Cannot Display This Video Mode" and then the information: "Optimum resolution 1280x1024 60Hz". I'm fairly new to using an operating system wholly via command line but I'm learning fast. :) I can access the server fine via SSH so using it is not a problem. However, for where it is positioned at the moment, it would be much more convenient to use the computer locally. I'm just wondering how to change the local screen resolution via SSH to something usable by my monitor. Thanks for your time! BTW: I have also tried switching displays and I still have the same problem. Also BTW: I tried putting in the xrandr command and all I get back is: Can't open display

    Read the article

  • Uninstalling a SQL Server Clustered Instance

    I have installed and uninstalled several instances of SQL Server in the past. Today, I need to uninstall a SQL Server 2008 R2 clustered instance. I have never uninstalled a clustered instance of SQL Server before. Can you provide a how-to guide to uninstall a clustered instance of SQL Server 2008 R2? NEW! SQL Monitor 2.0Monitor SQL Server Central's servers withRed Gate's new SQL Monitor.No installation required. Find out more.

    Read the article

  • SSRS Report Parts Versus Sub Reports FAQ

    I'm trying to decide on a development strategy to satisfy the reporting needs in my organization. I would like to increase our efficiency in responding to report requests, while minimizing our maintenance burden. Two topics that I would like to dig in to are Report Parts and Subreports. Can you provide some considerations for using one versus the other? NEW! SQL Monitor 2.0Monitor SQL Server Central's servers withRed Gate's new SQL Monitor.No installation required. Find out more.

    Read the article

  • Force Xorg to run in specific resolution

    - by Z9iT
    From my past experience (using Win-Xp), this particular monitor works only on 60Hz , Best resolution being 1024*768. I have "installed n configured" Ubuntu 12.04 Minimal (on USB stick) so that most of the time terminal is used, however whenever theirs a need to enter GUI I may issue startx command to go into gnome. However the problem is that on this particular system, issuing this command poses problem for its default refresh rate don't synchronize with the monitor. The display keep on flickering and utterly unreadable. It is visible that gnome has been loaded and default wallpaper and desktop items are visible. But the problem is due to refresh rate different than 60Hz. I am looking for a command attribute to startx command which will force the refresh rate to 60Hz and resolution preferably to 1204*768

    Read the article

  • Just installed 12.10 with WUBI and get error

    - by John
    Goes to the terminal screen and says: [drm:drm_crtc_helper_set_config] ERROR failed to set mode on 'crtc:10] From what I gather, this is an issue with my monitor. Not sure. I am really new to this. I have a Compaq Presario running XP SP3 and am using a Sony Bravia 32" LCD TV as a monitor. Downloaded Ubuntu from, well,... Ubuntu. Used WUBI to install. I suppose it created its own partition because install never asked. Is there a work-around for this? I have tried everything I know of, given my very limited experience. Thanks in advance.

    Read the article

  • Pillar Axiom OpenWorld Highlights

    - by uwes
    During the Storage General Session at Oracle OpenWorld Conference 2012 in San Francisco, the following Axiom-related announcements were made: Oracle Platinum Services for Axiom 600: Extending Oracle's Platinum Services to Axiom 600 as a standalone product –  the same level of service and support you get with Exadata – 24/7 fault monitoring, dedicated response and escalation management to meet enterprise-grade SLA’s, patch planning and management. Oracle Enterprise Manager Axiom Plug-in: Allowing DBA's to manage, maintain, monitor and provision the Axiom 600 storage system from Oracle EM. Oracle Virtual Machine Axiom Plug-in: Allowing Oracle VM and System Administrators to manage, maintain, monitor and provision the Axiom 600 storage system from Oracle VM using Storage Connect. Oracle Axiom Data Protection Manager 3.1: Leveraging Axiom's Copy Services, System Administrators can automatically create Application Consistent Clones of critical Windows and Oracle DataBase environments for quick recovery. For More Information Go To: Oracle.com Pillar Axiom Page Oracle Technology Network SAN Page

    Read the article

  • Changing Workspaces changes active window

    - by puk
    I have a dual monitor setup, but I am sure it's the same with a single monitor. If I have two applications open, one is vim, the other is google chrome. Lets say the focus is on vim. If I switch to another workspace (ie. alt3) then switch back to that workspace (ie. alt1) now the focus is on google chrome. This process toggles indefinitely so long as I don't transition too quickly. Is there any way to prevent/disable this?

    Read the article

  • To check the default select value tag!

    - by Small Wolf
    I have a question.Here is the code! f.select(:departments,Department.all.collect{|c|[c.name,c.id]},{},:size=>10,:multiple => ture) class Emergency has many :departments end the html source like this: <select id="emergency_departments" multiple="multiple" name="emergency[departments][]" size="10"><option value ="">""</option>....</select> now I want to get the default selected tag , who can tell me how ?

    Read the article

  • enable iptables firewall on linux

    - by user13278061
     Here is a very basic set of instruction to setup a simple iptables firewall configuration on linux (redhat) Enable firewall log as root thenenter the following command, it launch a text gui #> setup first screen: Choose firewall configuration second screen: choose "Enabled" then "Customize" third screen: select you interface in "Trusted Devices", select "Allow Incoming" for "SSH" "Telnet" "FTP" (add eventually other ports, then press "OK" (2 times, then "Quit") At that point the firewall is enabled. You can start/stop/monitor using service iptables start/stop/status Change timeout to changed the tcp established connection timeout #> echo 120 >    /proc/sys/net/ipv4/netfilter/ip_conntrack_tcp_timeout_established Monitor connection in iptables tables for example if you want to track a connection establish from a host  152.68.65.207 #> cat /proc/net/ip_conntrack |grep 152.68.65.207

    Read the article

  • javax.crypto AES encryption - Do I only need to call doFinal?

    - by Spines
    I want to do AES CBC encryption in Java. I'm using javax.crypto. After I have the Cipher initialized, do I only need to call doFinal on the clear bytes to properly encrypt it? Or do I need to do something with update? Documentation says update: Continues a multiple-part encryption or decryption operation and doFinal Encrypts or decrypts data in a single-part operation, or finishes a multiple-part operation what exactly do they mean by multiple-part encryption?

    Read the article

  • Hibernate Query - Get latest versions by timestamp?

    - by Slim
    I have a database that is being used as a sort of version control system. That is, instead of ever updating any rows, I add a new row with the same information. Each row also contains a version column that is a date timestamp, so the only difference is the new row will have a more recent timestamp. What I'm having trouble with is writing an efficient hibernate query to return the latest version of these rows. For the sake of example, these are rows in a table called Product, the timestamped column is version. There are multiple versions of multiple products in the table. So there may be multiple versions (rows) of ProductA, multiple versions of ProductB, etc. And I would like to grab the latest version of each. Can I do this in just a single hibernate query? session.createQuery("select product from Product product where...?"); Or would this require some intermediate steps?

    Read the article

  • 12.04 + Alienware M11x R2 + Bumblebee 3.0 = low resolution only

    - by user89171
    I had a fresh install of Ubuntu 12.04 on my Alienware M11x, and it worked with the native monitor resolution of 1366x768. In trying to get the Optimus chipset working, I installed Bumblebee 3.0. Now, I am only offered 640x480 for my monitor resolution. Graphically, Unity 3D appears to be working now, instead of the Unity 2D I had before, so something went right, but I don't know how I can get it to offer me any higher resolutions than 640x480. I've looked up many pages that address this topic, but nothing that I've seen suggested has worked. sudo add-apt-repository ppa:bumblebee/stable sudo apt-get update sudo apt-get install bumblebee-nvidia was the latest thing I tried. I've tried some variants of this, I've been sure to uninstall nvidia drivers prior to reinstalling Bumblebee and various video drivers. Does anyone have any clue as to how I can get back to native resolution?

    Read the article

  • Handling Backups for Rapid Resilience

    The backup and restore system in SQL Server hasn't changed a great deal over the years despite a huge growth in the typical size of databases. When disaster strikes, and an important service is taken offline while a restore is performed, there is often time to reflect on whether it might be possible to design databases for a more rapid recovery of the most critical parts of a database application. Uncover the flexibility of custom metricsKeep up to date from anywhere with SQL Monitor, and monitor the most important data for your servers and applications. Download a free trial.

    Read the article

  • Can I play any Buffer only once at a given time?

    - by mystify
    From the OpenAL documentation: The basic OpenAL objects are a Listener, a Source, and a Buffer. There can be a large number of Buffers, which contain audio data. Each buffer can be attached to one or more Sources My problem is, that I have one sound file which I need to play multiple times per second, at the same time. The sound is 2 seconds long. So it will overlap. Would I need multiple filled buffers for this (= multiple times that sound in memory)? If I would attach one Buffer to multiple Sources, would I be able to play the sound 10 times, overlapping itself, with just one copy in memory? Or would I still have to deal with 10 copies of that sound in memory?

    Read the article

  • How to get warning massage when clicking close(X) button of browser with tabs in Google Crome

    - by Shantanu
    How to get a warning massage if i accidently click the close button of the chrome browser which have multiple tabs opened at that time? I am a regular user of Crome and having this problem while using it. I normally open multiple tabs inside a single browser but sometimes i accidently click the close button of browser and as soon as the button is clicked crome does't give any warning issue about multiple active tabs and close the entire window. Is the end user like me is browsing on normal crome window then he can open the websites again by checking the web history but if he is browsing inside private browser then he can't do anything(this happens with me very regularly because i normally browse in private browser). On the other hand if you accidently click the close button in mozilla which have multiple tabs open it throws a warning massage to the user and asks for his wish.

    Read the article

  • Distributed and/or Parallel SSIS processing

    - by Jeff
    Background: Our company hosts SaaS DSS applications, where clients provide us data Daily and/or Weekly, which we process & merge into their existing database. During business hours, load in the servers are pretty minimal as it's mostly users running simple pre-defined queries via the website, or running drill-through reports that mostly hit the SSAS OLAP cube. I manage the IT Operations Team, and so far this has presented an interesting "scaling" issue for us. For our daily-refreshed clients, the server is only "busy" for about 4-6 hrs at night. For our weekly-refresh clients, the server is only "busy" for maybe 8-10 hrs per week! We've done our best to use some simple methods of distributing the load by spreading the daily clients evenly among the servers such that we're not trying to process daily clients back-to-back over night. But long-term this scaling strategy creates two notable issues. First, it's going to consume a pretty immense amount of hardware that sits idle for large periods of time. Second, it takes significant Production Support over-head to basically "schedule" the ETL such that they don't over-lap, and move clients/schedules around if they out-grow the resources on a particular server or allocated time-slot. As the title would imply, one option we've tried is running multiple SSIS packages in parallel, but in most cases this has yielded VERY inconsistent results. The most common failures are DTExec, SQL, and SSAS fighting for physical memory and throwing out-of-memory errors, and ETLs running 3,4,5x longer than expected. So from my practical experience thus far, it seems like running multiple ETL packages on the same hardware isn't a good idea, but I can't be the first person that doesn't want to scale multiple ETLs around manual scheduling, and sequential processing. One option we've considered is virtualizing the servers, which obviously doesn't give you any additional resources, but moves the resource contention onto the hypervisor, which (from my experience) seems to manage simultaneous CPU/RAM/Disk I/O a little more gracefully than letting DTExec, SQL, and SSAS battle it out within Windows. Question to the forum: So my question to the forum is, are we missing something obvious here? Are there tools out there that can help manage running multiple SSIS packages on the same hardware? Would it be more "efficient" in terms of parallel execution if instead of running DTExec, SQL, and SSAS same machine (with every machine running that configuration), we run in pairs of three machines with SSIS running on one machine, SQL on another, and SSAS on a third? Obviously that would only make sense if we could process more than the three ETL we were able to process on the machine independently. Another option we've considered is completely re-architecting our SSIS package to have one "master" package for all clients that attempts to intelligently chose a server based off how "busy" it already is in terms of CPU/Memory/Disk utilization, but that would be a herculean effort, and seems like we're trying to reinvent something that you would think someone would sell (although I haven't had any luck finding it). So in summary, are we missing an obvious solution for this, and does anyone know if any tools (for free or for purchase, doesn't matter) that facilitate running multiple SSIS ETL packages in parallel and on multiple servers? (What I would call a "queue & node based" system, but that's not an official term). Ultimately VMWare's Distributed Resource Scheduler addresses this as you simply run a consistent number of clients per VM that you know will never conflict scheduleing-wise, then leave it up to VMWare to move the VMs around to balance out hardware usage. I'm definitely not against using VMWare to do this, but since we're a 100% Microsoft app stack, it seems like -someone- out there would have solved this problem at the application layer instead of the hypervisor layer by checking on resource utilization at the OS, SQL, SSAS levels. I'm open to ANY discussion on this, and remember no suggestion is too crazy or radical! :-) Right now, VMWare is the only option we've found to get away from "manually" balancing our resources, so any suggestions that leave us on a pure Microsoft stack would be great. Thanks guys, Jeff

    Read the article

  • WMI permissions: Select CommandLine, ProcessId FROM Win32_Process returns no data for CommandLine

    - by user57935
    Hi all, I am gathering performance data via WMI and would like to avoid having to use an account in the Administrators group for this purpose. The target machine is running Windows Server 2003 with the latest SP/updates. I've done what I believe to be the appropriate configuration to allow our user access to WMI (similar to what is described here: http://msdn.microsoft.com/en-us/library/aa393266.aspx). Here are the specific steps that were followed: Open Administrative Tools - Computer Management: Under Computer Management (Local) Expand Services and Applications, right click WMI Control and select properties. In the Security tab, expand Root, highlight CIMV2, click Security (near bottom of window); add Performance Monitor Users and enable the options : Enable Account and Remote Enable. ­Open Administrative Tools - Component Services: Under Console Root go to Component Services- Computers - Right click My Computer and select properties, select the COM security tab, in “Access Permissions” click "Edit Default" select(or add then select) “Performance Monitor Users” group and allow local access and remote access and click ok. In “Launch and Activation Permissions” click “Edit Default” select(or add then select) “Performance Monitor Users” group and allow Local and Remote Launch and Activation Permissions. ­Open Administrative Tools - Component Services: Under Console Root go to Component Services- Computers - My Computer - DCOM Config - highlight “Windows Management and Instrumentation” right click and select properties, Select the Security tab, Under “Launch and Activation Permissions” select Customize, then click edit, add the “Performance Users Group” and allow local and remote Remote Launch and Remote Activation privileges. I am able to connect remotely via WMI Explorer but when I perform this query: Select CommandLine, ProcessId FROM Win32_Process I get a valid result but every row has an empty CommandLine. If I add the user to the Administrators group and re-run the query, the CommandLine column contains the expected data. It seems there is a permission I am missing somewhere but I am not having much luck tracking it down. Many thanks in advance.

    Read the article

  • WMI permissions: Select CommandLine, ProcessId FROM Win32_Process returns no data for CommandLine

    - by user57935
    I am gathering performance data via WMI and would like to avoid having to use an account in the Administrators group for this purpose. The target machine is running Windows Server 2003 with the latest SP/updates. I've done what I believe to be the appropriate configuration to allow our user access to WMI (similar to what is described here: http://msdn.microsoft.com/en-us/library/aa393266.aspx). Here are the specific steps that were followed: Open Administrative Tools - Computer Management: Under Computer Management (Local) Expand Services and Applications, right click WMI Control and select properties. In the Security tab, expand Root, highlight CIMV2, click Security (near bottom of window); add Performance Monitor Users and enable the options : Enable Account and Remote Enable. ­Open Administrative Tools - Component Services: Under Console Root go to Component Services- Computers - Right click My Computer and select properties, select the COM security tab, in “Access Permissions” click "Edit Default" select(or add then select) “Performance Monitor Users” group and allow local access and remote access and click ok. In “Launch and Activation Permissions” click “Edit Default” select(or add then select) “Performance Monitor Users” group and allow Local and Remote Launch and Activation Permissions. ­Open Administrative Tools - Component Services: Under Console Root go to Component Services- Computers - My Computer - DCOM Config - highlight “Windows Management and Instrumentation” right click and select properties, Select the Security tab, Under “Launch and Activation Permissions” select Customize, then click edit, add the “Performance Users Group” and allow local and remote Remote Launch and Remote Activation privileges. I am able to connect remotely via WMI Explorer but when I perform this query: Select CommandLine, ProcessId FROM Win32_Process I get a valid result but every row has an empty CommandLine. If I add the user to the Administrators group and re-run the query, the CommandLine column contains the expected data. It seems there is a permission I am missing somewhere but I am not having much luck tracking it down. Many thanks in advance.

    Read the article

  • What's hogging my CPU?

    - by endolith
    Ubuntu's System Monitor applet shows 100% CPU usage continuously. If I click it, the resources tab shows it at 100% continuously, too. If I go to processes, though, to find out which process is the culprit, there is nothing above 10%. If I run top there is nothing above 10%. I try killing lots of things, but it continues at 100%. How can I find out what's hogging the CPU? This is an unusual situation on a computer I use daily, that normally only hits 100% CPU when I'm doing something that requires it (like loading 32 Firefox tabs) after which it goes back to a normal idle level. It's not a new install or anything. It shouldn't be maxed out. I'm not sure when it started or if I changed something that caused it to happen. Normally I would use top or System Monitor and find the process that had gone out of control, but I can't find anything with those tools this time. It persists after reboots and everything. And the processor is obviously hot, so it's not an erroneous reading. Update: I tried killing any process I saw active again, and killing vino-server finally fixed the problem, even though it never went above 5%. I had enabled Remote Desktop a few days ago (and have obviously now disabled it). How did it manage to use 100% CPU while top only showed it as 5% or so? How do I identify the culprit in the future? Looks like I'm not the only one: Still a problem in both jaunty & karmic. Interestingly, both System Monitor & htop do not show the sum of individual processes being anywhere near 100% cpu.

    Read the article

  • Display stretches 4:3 ratios; Adds scrolling to other ratios

    - by Matt
    I have a dual monitor setup. Normally, they both display at 1680x1050. They have been setup this way for about a year. I'm using Windows XP Professional 2003 x64 SP2. Today, out of nowhere, one of the monitors kicked back to a lower resolution. I was not playing with any configuration at the time.. in fact all I had done was close a window (maybe a browser). But the thing is that the resolution is still preserved partially by the fact that the screen will scroll when you move the mouse. So it's like looking through a 1024x768 window into a 1680x1050 world. The monitor itself does not appear to be damaged, because I also have it connected to my netbook (via KVM) and higher resolutions work fine. I tried uninstalling/reinstalling the drivers to no avail. System restore doesn't help either. I'm unsure of the exact ATI card I'm using.. Device Manager lists it as "Radeon X300/X550/X1050". There is no Catalyst Control Center software installed. I tried to install it, but there doesn't seem to be a way to install it by itself ... it forces you to install another driver, which breaks both of my displays, forcing me to go into safe mode and run system restore again. Any ideas? Thanks EDIT: After playing around more, I discovered that the "scrolling" behavior is only present for aspect ratios that are not 4:3. For 4:3 ratios, it just stretches out to fit the wide screen. My monitor's native ratio is 16:9 .. what could be causing it to think it needs to scroll?

    Read the article

  • Ubuntu 13.10 on Acer V5-472 with HD 4000

    - by Hyperboreus
    I have an Acer V5-472 with intel HD4000 graphics chipset and a built-in 1366*768 display. I have installed ubuntu 13.10 amd64 in legacy boot mode with an external monitor. Installation showed no problems, I can boot from HDD and log into my system. The internal display doesn't work and I have to use an external monitor. I have tried the following (found in other threads) to no avail: Setting grub option "acpi_backlight=vendor" or "acpi_osi=Linux" or both. Installing the intel HD drivers for Linux from their homepage. Running in circles, screaming and shouting. The internal display lights up (I can change the brightness with Fn-Left and Fn-Right) but that's all. When I boot, I get a purple splash screen and from then only the external monitor works. I read somewhere that this might be a problem with kernel 3.11? Has anybody ubuntu running on an Acer V5-472? Should I change ubuntu version or use 32-bit instead? In general, how can I get the internal display to work? Edit: The settings-display dialogue shows the internal display correctly with supported resolution of 1366.

    Read the article

  • Windows 7 blank dvi screen

    - by user99
    I've just upgraded to Windows 7 Ultimate 64 bit. I had an issue during installation; after setup rebooted instead of going to the 'Completing installation' screen I just got a blank screen. I eventually(!) figured out that this meant 'un-plug your second monitor to proceed'. When I did this, installation completed in a snap and everything runs fine. However, whenever I plug my second monitor into my PC, it gets no signal, the the primary monitor removes all windows. icons, the taskbar and the cursor just shows the desktop wallpaper. I'm running a GTS8800 512mb, with the latest drivers (197 IIRC). The monitors are identical, and both plug into DVI sockets on the graphics card, the only difference is I connect one using a straight DVI cable and the other using a VGA cable and a VGA-DVI converter. It's the DVI cabled one that has the issues (if I plug it in by itself it gets no signal). Everything was working fine before I upgraded to Windows 7 (I used to run XPSP2). Anyone have any ideas?

    Read the article

  • Building a PC for Work and play? [closed]

    - by Derek Organ
    Ok, Its been a long time since I build my own PC so I'm looking to get back into it again and build a new one. First off budget is about €800 excluding the monitor and windows 7 licence and mouse. (just bought a new g500) I plan on using my computer for work, lots of applications open at once but none particularly excessive (photoshop being the most demanding, mostly coding tools) I also use it for some gaming, e.g. COD, Starcraft etc. One thing I do want to do eventually is get a really good monitor with hight resolution and maybe 27" so the graphics card needs to be able to make best use of that. So a few questions 1) Is the bottle neck in performance mostly still the harddrives? 2) Aren't most processors e.g. i5 etc even i3 so far a head of other bottlenecks it makes litte difference the higher you go. Isn't the Graphics card dealing with heavy graphics so what really slows because of a slow CPU? So from this my thinking is to get a SSD drive as my primary drive for OS etc and have loads of memory e.g. 6-8GB and a decent mid level graphics card? It doesn't seem at my level worth spending much on CPU and any other parts really. I basic parts off the top of my head Case, Motherboard CPU SSD Drive SATA Drive Power Supply Memory Cooling (fan?) Graphics Card Network Card Keyboard DVD drive Mouse Windows Monitor Am I missing anything? Any helpful tips or general education much appreciated. Thanks, Derek

    Read the article

< Previous Page | 346 347 348 349 350 351 352 353 354 355 356 357  | Next Page >