Search Results

Search found 8373 results on 335 pages for 'hardware recommendation'.

Page 142/335 | < Previous Page | 138 139 140 141 142 143 144 145 146 147 148 149  | Next Page >

  • When using RAID10 + BBWC why is it better to separate PostgreSQL data files from OS and transaction logs than to keep them all on the same array?

    - by Vlad
    I've seen the advice everywhere (including here and here): keep your OS partition, DB data files and DB transaction logs on separate discs/arrays. The general recommendation is to use RAID1 for OS, RAID10 for data (or RAID5 if load is very read-biased) and RAID1 for transaction logs. However, considering that you will need at least 6 or 8 drives to build this setup, wouldn't a RAID10 over 6-8 drives with BBWC perform better? What if the drives are SSDs? I'm talking here about internal server drives, not SAN.

    Read the article

  • can't get past the login screen

    - by Greg
    Using a brand-new install to a usb stick of 12.04 lts installed by Universal USB Installer 1.8.9.8. I log in as "ubuntu" with a blank password, the console appears for a second or two with text scrolling past and then it returns to the login page. I've used the same usb stick on several computers with the same results, so it doesn't appear to be a hardware/driver issue. I have not tried installing to the hard drive, because I wanted to try it out first.

    Read the article

  • Is there a simple, flat, XML-based query-able data storage solution? [closed]

    - by alex gray
    I have been in long pursuit of an XML-based query-able data store, and despite continued searches and evaluations, I have yet to find a solution that meets the my needs, which include: Data is wholly contained within XML nodes, in flat text files. There is a "native" - or at least unobtrusive - method with which to perform Create/Read/Update/Delete (CRUD) operations onto the "schema". I would consider access via http, XHR, javascript, PHP, BASH, or PERL to be unobtrusive, dependent on the complexity of the set of dependencies. Server-side file-system reads and writes. A client-side interface element, accessible in any browser without a plug-in. Some extra, preferred (but optional) requirements include: Respond to simple SQL, or similarly syntax queries. Serve the data on a bare bones https server, with no "extra stuff", either via XMLHTTPRequest, HTTP proper, or JSON. A few thoughts: What I'm looking for may be possible via some Java server implementations, but for the sake of this question, please do not suggest that - unless it meets ALL the requirements. Java, especially on the client-side is not really an option, nor is it appealing from a development viewpoint.* I know walking the filesystem is a stretch, and I've heard it's possible with XPATH or XSLT, but as far as I know, that's not ready for primetime, nor even yet a recommendation. However the ability to recursively traverse the filesystem is needed for such a system to be of useful facility. At this point, I have basically implemented what I described via, of all things, CGI and Bash, but there has to be an easier way. Thoughts?

    Read the article

  • Ubuntu Server driver source catalog

    - by SRU IT
    Where do I find this drivers on my .iso image for the server software, or what is the ftp: location to download these files as those are the two options i am given on how to solve this issue. ubuntu-server 11.10, hardware: dell r210 what I am reading is that I am possibly missing 'mpt2sas' driver somewhere in this boot disc. I'm attempting to download a different version in the hopes that the mistake I made was downloading software meant for an AMD architecture.

    Read the article

  • FOUR questions to ask if you are implementing DATABASE-AS-A-SERVICE

    - by Sudip Datta
    During my ongoing tenure at Oracle, I have met all types of DBAs. Happy DBAs, unhappy DBAs, proud DBAs, risk-loving DBAs, cautious DBAs. These days, as Database-as-a-Service (DBaaS) becomes more mainstream, I find some complacent DBAs who are basking in their achievement of having implemented DBaaS. Some others, however, are not that happy. They grudgingly complain that they did not have much of a say in the implementation, they simply had to follow what their cloud architects (mostly infrastructure admins) offered them. In most cases it would be a database wrapped inside a VM that would be labeled as “Database as a Service”. In other cases, it would be existing brute-force automation simply exposed in a portal. As much as I think that there is more to DBaaS than those approaches and often get tempted to propose Enterprise Manager 12c, I try to be objective. Neither do I want to dampen the spirit of the happy ones, nor do I want to stoke the pain of the unhappy ones. As I mentioned in my previous post, I don’t deny vanilla automation could be useful. I like virtualization too for what it has helped us accomplish in terms of resource management, but we need to scrutinize its merit on a case-by-case basis and apply it meaningfully. For DBAs who either claim to have implemented DBaaS or are planning to do so, I simply want to provide four key questions to ponder about: 1. Does it make life easier for your end users? Database-as-a-Service can have several types of end users. Junior DBAs, QA Engineers, Developers- each having their own skillset. The objective of DBaaS is to make their life simple, so that they can focus on their core responsibilities without having to worry about additional stuff. For example, if you are a Developer using Oracle Application Express (APEX), you want to deal with schema, objects and PL/SQL code and not with datafiles or listener configuration. If you are a QA Engineer needing database copies for functional testing, you do not want to deal with underlying operating system patching and compliance issues. The question to ask, therefore, is, whether DBaaS makes life easier for those users. It is often convenient to give them VM shells to deal with a la Amazon EC2 IaaS, but is that what they really want? Is it a productive use of a developer's time if he needs to apply RPM errata to his Linux operating system. Asking him to keep the underlying operating system current is like making a guest responsible for a restaurant's decor. 2. Does it make life easier for your administrators? Cloud, in general, is supposed to free administrators from attending to mundane tasks like provisioning services for every single end user request. It is supposed to enable a readily consumable platform and enforce standardization in the process. For example, if a Service Catalog exposes DBaaS of specific database versions and configurations, it, by its very nature, enforces certain discipline and standardization within the IT environment. What if, instead of specific database configurations, cloud allowed each end user to create databases of their liking resulting in hundreds of version and patch levels and thousands of individual databases. Therefore the right question to ask is whether the unwanted consequence of DBaaS is OS and database sprawl. And if so, who is responsible for tracking them, backing them up, administering them? Studies have shown that these administrative overheads increase exponentially with new targets, and it could result in a management nightmare. That leads us to our next question. 3. Does it satisfy your Security Officers and Compliance Auditors? Compliance Auditors need to know who did what and when. They also want the cloud platform to be secure, so that end users have little freedom in tampering with it. Dealing with VM sprawl is not the easiest of challenges, let alone dealing with them as they keep getting reconfigured and moved around. This leads to the proverbial needle in the haystack problem, and all it needs is one needle to cause a serious compliance issue in the enterprise. Bottomline is, flexibility and agility should not come at the expense of compliance and it is very important to get the balance right. Can we have security and isolation without creating compliance challenges? Instead of a ‘one size fits all approach’ i.e. OS level isolation, can we think smartly about database isolation or schema based isolation? This is where the appropriate resource modeling needs to be applied. The usual systems management vendors out there with heterogeneous common-denominator approach have compromised on these semantics. If you follow Enterprise Manager’s DBaaS solution, you will see that we have considered different models, not precluding virtualization, for different customer use cases. The judgment to use virtual assemblies versus databases on physical RAC versus Schema-as-a-Service in a single database, should be governed by the need of the applications and not by putting compliance considerations in the backburner. 4. Does it satisfy your CIO? Finally, does it satisfy your higher ups? As the sponsor of cloud initiative, the CIO is expected to lead an IT transformation project, not merely a run-of-the-mill IT operations. Simply virtualizing server resources and delivering them through self-service is a good start, but hardly transformational. CIOs may appreciate the instant benefit from server consolidation, but studies have revealed that the ROI from consolidation would flatten out at 20-25%. The question would be: what next? As we go higher up in the stack, the need to virtualize, segregate and optimize shifts to those layers that are more palpable to the business users. As Sushil Kumar noted in his blog post, " the most important thing to note here is the enterprise private cloud is not just an IT project, rather it is a business initiative to create an IT setup that is more aligned with the needs of today's dynamic and highly competitive business environment." Business users could not care less about infrastructure consolidation or virtualization - they care about business agility and service level assurance. Last but not the least, lot of CIOs get miffed if we ask them to throw away their existing hardware investments for implementing DBaaS. In Oracle, we always emphasize on freedom of choosing a platform; hence Enterprise Manager’s DBaaS solution is platform neutral. It can work on any Operating System (that the agent is certified on) Oracle’s hardware as well as 3rd party hardware. As a parting note, I urge you to remember these 4 questions. Remember that your satisfaction as an implementer lies in the satisfaction of others.

    Read the article

  • New sales kit for partners: Oracle Enterprise Manager 12c

    - by Javier Puerta
    Check out the latest Quick Reference Guides for Enterprise Manager 12c in the Knowledge Zone. The two-page Quick Reference Guide is designed to help partners uncover additional revenue opportunity by positioning Enterprise Manager in your sales engagement. Content includes elevator pitch for Enterprise Manager, tips on identifying target customers, qualifying questions to initiate customers discussion, supporting videos, references, and whitepapers for each customer scenario: Enterprise Manager 12c for Application Partners Enterprise Manager 12c for Hardware Partners Enterprise Manager 12c for Database Partners

    Read the article

  • Nvidia Proprietary Drivers

    - by Ben
    I have an X79 motherboard and a GTX 670. I have tried to install every combination of Ubuntu 11.10, 12.04, 12.10, Linux Mint 13 and the Nvidia Driver 304.43, 295.xx, etc. I have read everything and pretty much tried everything...drivers from the site, PPAs, some random scripts, editing the xorg.conf file. It has been four days of futile effort. Should I just wait...or something? Has anyone gotten hardware like this to work?

    Read the article

  • KWin Can Cause A Performance Hit Too, But It's Different From Compiz

    <b>Phoronix:</b> "This surprised many so we published another article entitled The Cost Of Running Compiz where we showed the performance penalties of a compositing window manager with different hardware and drivers. This led some to ask whether the performance of KWin also causes the OpenGL frame-rate to drop, so here are those KDE benchmarks."

    Read the article

  • search nfs network volume from mac client

    - by user1440190
    Asked: how does a Mac OSX SL or Lion user search the cluster for a particular file (foo.txt) "From the cluster, you would need to run some form of recursive lookup for the file desired. As an example, using 'find'. RAM-1# find /ifs |grep test.txt /ifs/Elements/avid2test.txt /ifs/Elements/test.txt I would suggest contacting Apple support regarding their recommendation for searching for files on remote file systems from the Mac client itself" OK that's great, but I don't want users using CLI !!! Anyone know a good method (non-CLI)? Spotlight is not an option. BTW cluster is roughly 80TB

    Read the article

  • No wired and wireless connection found after upgrading 12.04

    - by ag108
    After upgrading 12.04 LTS, my wired connection works but no wireless connection found. I tried many suggestion for wireless connection on Ubuntu forum but nothing works for me. Now, maybe by mistake, I am also unable to connect Internet through wired connection. Now, (1) its says "No network devices available" and (2) in System Settings ? Hardware ? no additional driver icon is showing Please help. I don't know much about Ubuntu.

    Read the article

  • Which of these courses are the hardest and why?

    - by DSL Client
    Which of these courses are the hardest and why? What should I watch out for? Probability and Statistics for Computer Science Introduction to Software Engineering Data Structures and Algorithms Operating Systems Introduction to Theoretical Computer Science System Hardware Advanced Program Design with C++ Information Systems Security Computer Architecture Databases Web Programming Computer Graphics Digital System Design

    Read the article

  • IIS 7 SSO stops working during high CPU load? [migrated]

    - by DanB
    On our IIS7 site (Windows 2008 Server), we have set up single sign-on (SSO). It seems to work fine most of the time, but when the CPU load becomes high, SSO authentication completely stops working. I did some research and tried this suggestion to increase the max number of worker processes in the default app pool, but the increase did not help. Some details: The site is a WordPress blog. The server has plenty of RAM (2 GB) and free disk space. SSO is achieved by putting a copy of the WordPress login page (wp-login.php) into a subfolder below the root that has anonymous authentication disabled, and then redirecting the browser to it. This was the recommendation of Microsoft given to our consultants. To increase CPU load for testing, I have three scripts hit the home page simultaneously, over and over. This drives CPU to 100%. When these scripts are running, SSO authentication simply doesn't happen. As soon as I stop the scripts, SSO works again. (I should mention that the SSO problem also happens when many users visit the site at once....) The WordPress database process (mysqld) is not stressed at all by the scripts. I would be happy to provide further diagnostics. Any help appreciated!

    Read the article

  • NFS of NAS server blocks in cluster environment

    - by Zardoz
    In our department we have an Iomega NAS (px4-300d) connected to a Supermicro cluster with 5 nodes (12 cores per node). Each node mounts a share on that NAS by using NFS. Unfortunately after some time (several minutes) of permanent read/write operations (from all nodes) the NAS starts to block and a bit later freezes completely. We tried several options of the mount command, but nothing helped (async, intr, wsize, rsize). The NAS itself doesn't allow many options (better to say none). Do you have any recommendation how to integrate a NAS using NFS in a cluster environment?

    Read the article

  • Can't read from the source file or disk

    - by Wanna coffee
    I'm having a two WD external hard disk with capacity of 1 TB. I'm trying to copy SAP file(capacity - 250 GB ) in the extension of .vmdk from one hard disk to another hard disk. But when ever i'm trying to copy, at down to the line it showing me this error message. By default my both hard disk File System value is NFTS, even though it showing me an this error message. Is this problem with OS or Hard disk or Data which i'm taken into the action?? What might be the problem, Please give me your suggestions and recommendation. Awaiting for your reply.

    Read the article

  • Kubuntu 11.10 Lot of Networking problems

    - by Cobraone
    Since I upgrade to 11.10 I have a lot of problems with KDE. First of all there are problems in configuring a static IP address. Just to explain @home I have a normal fiber ADSL and I use a DHCP. When I go to a customer I must insert a static IP address. With ifconfig everything seems ok but there is something wrong in searching DNS names. (I have installed Ubuntu and was going ok again). Now I Have reinstalled again Kubuntu 11.10 and I have the same problem in addition today I have discovered that if I connect to a network in another customer office the desktop freezes and I could only switch between windows with alt+tab. No FN key or right click to open run command works. So i unplugged network (configuration is just DHCP here) and tried on another position in office. It was the same. My Laptop freezes when connected, a fedora 14 of a friend works. So I decided to connect my Galaxy S II as USB network device. Everything is ok for like 3 minutes. When I noticed a little loss of signal again the desktop freezes and i must work (like now) just switching between windows with alt+tab). Additional information: Unplugging network or restarting it via Konsole does not not solve the freezing problem. Every time I must open a console and reboot. Any idea of what tests to do ? Just a recommendation: If I must post here logs or something else please guide me. I use Linux since Ubuntu 9 but I am not an "expert".

    Read the article

  • Tune your Linux kernel with sysctl

    <b>Ghacks:</b> "Today most kernels work out of the box with most hardware. But there are times when you could squeeze out a bit more performance or even lighten your kernel on the fly. You can do that with the sysctl command."

    Read the article

  • More Tables or More Databases?

    - by BuckWoody
    I got an e-mail from someone that has an interesting situation. He has 15,000 customers, and he asks if he should have a database for their data per customer. Without a LOT more data it’s impossible to say, of course, but there are some general concepts to keep in mind. Whenever you’re segmenting data, it’s all about boundary choices. You have not only boundaries around how big the data will get, but things like how many objects (tables, stored procedures and so on) that will be involved, if there are any cross-sections of data (do they share location or product information) and – very important – what are the security requirements? From the answer to these types of questions, you now have the choice of making multiple tables in a single database, or using multiple databases. A database carries some overhead – it needs a certain amount of memory for locking and so on. But it has a very clean boundary – everything from objects to security can be kept apart. Having multiple users in the same database is possible as well, using things like a Schema. But keeping 15,000 schemas can be challenging as well. My recommendation in complex situations like this is similar to a post on decisions that I did earlier – I lay out the choices on a spreadsheet in rows, and then my requirements at the top in the columns. I  give each choice a number based on how well it meets each requirement. At the end, the highest number wins. And many times it’s a mix – perhaps this person could segment customers into larger regions or districts or products, in a database. Within that database might be multiple schemas for the customers. Of course, he needs to query across all customers, that becomes another requirement. Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • Ubuntu 13.10 - Wifi not re-connecting after suspend

    - by Matthew
    I upgraded the other day from Ubuntu 13.04 to 13.10 on my ZaReason laptop, and the WiFi doesn't reconnect after suspend. In 13.04, pressing Fn+F2 turned network connections back on after suspend. The 'soft' wireless hardware switch. But in 13.10, that doesn't work, and trying to enable networking in the networking menu doesn't work either. So I have to restart every time. Any suggestions to make it work?

    Read the article

  • Performance Tune IBM DB2 z/OS Applications using Resource Constraint Analysis

    For the DB2 for z/OS professional the two most common systems tuning scenarios are tuning a DB2 data sharing group or tuning a series of application SQL statements. The data sharing group environment can involve multiple hardware installations and many other cross-system features and functions such as coupling facilities and management policies. Resource constraint analysis is a useful tool in both situations.

    Read the article

  • New Sales Kit – Enterprise Manager 12c

    - by Cinzia Mascanzoni
    Check out the latest Quick Reference Guides for Enterprise Manager 12c in the Knowledge Zone. The two-page Quick Reference Guide is designed to help partners uncover additional revenue opportunity by positioning Enterprise Manager. Content includes elevator pitch for Enterprise Manager, tips on identifying target customers, qualifying questions to initiate customers discussion, supporting videos, references, and whitepapers for each customer scenario.• Enterprise Manager 12c for Application Partners • Enterprise Manager 12c for Hardware Partners• Enterprise Manager 12c for Database Partners

    Read the article

  • Which universal or driverless printing solution do you use/recommend?

    - by Matt
    I'm in need of a driverless printing solution for Microsoft Terminal Services 2003/2008. This is mainly to support clients who are connected through broadband into our hosted servers. We were hoping that MSTS 2008 thinprint would be the answer but unfortunately it performs poorly in the print area. The files are too large. I found the following slightly outdated URL: http://www.msterminalservices.org/software/Printing/ This lists a number of products but I have no experience with any of them. I'd like a product that works/easy to install (as our clients are remote and not particularly tech savvy) and ideally I just pay for the server license and not every clients. What is your experience/recommendation and tips you can offer me in regards to TS printing? thanks in advance.

    Read the article

  • Ubuntu Server Configuration -- Harddrive Partitioning

    - by black_bird
    Currently Ubuntu Server is telling me that when I'm making a new partition for Ubuntu Server on this NTFS 1TB HD that I currently have installed to the hardware, that the partition must be a minimum of 52% of the hard drive space or ~521GB. I'm almost positive that this will run into other data, as I have quite a bit of stuff on the hard drive currently. Can I not make a Ubuntu server partition on that hard drive at like 100GB or something? Why does it require so much?

    Read the article

  • Les WebAPIs de Mozilla progressent, les applications Web peuvent exploiter de plus en plus de fonctionnalités natives des Smartphones

    Les WebAPIs de Mozilla progressent Et permettent aux applications Web d'exploiter de plus en plus de fonctionnalités du hardware des smartphones Les frontières s'effacent entre le développement mobile et le Web. Le mouvement a commencé avec des applications hébergées pour contourner les restrictions des galeries d'applications (lire par ailleurs). Le trio HTML5, CSS3, JavaScript a fait de plus en plus d'adepte à tel point que certains lui prédisent même un avenir plus florissant que les développements natifs. La Fondation Mozilla oeuvre à sa...

    Read the article

  • Virtualized Development Server for simulating 3-Tier Environment

    - by chris.cyvas
    Hello, I am thinking about buying a new server based development box for development (redundantly redundant, I know ;)). Ideally, I want to run something like ESXi or Xen Hypervisor at the lowest level. Then I want to add (at least) 5 Linux VM's for the following uses: 2 Web Servers 2 Application Servers 1 Database Server I want to load balance the 2 web servers and the 2 application servers and (somewhat obviously) they need to be all networked together to simulate a production environment. Also, it used to be the case that the recommendation was to put each VM on it's own hard drive, but I'm not sure that holds water anymore. Any advice? Does anyone have any advice on how to pull this off? Gotchya's, LookOuts!, etc? Thanks!

    Read the article

< Previous Page | 138 139 140 141 142 143 144 145 146 147 148 149  | Next Page >