Search Results

Search found 21551 results on 863 pages for 'company questions'.

Page 524/863 | < Previous Page | 520 521 522 523 524 525 526 527 528 529 530 531  | Next Page >

  • How can I learn more about ADF?

    - by jhpierce -Oracle
    Look to the Oracle Technology Network for a wealth of information, tutorials, best practices and coding examples. The place to start is the Oracle Application Development Framework (ADF) web page. The Oracle ADF page has basic information and downloads for ADF, but the real wealth is in the links to other pages. One of the pages is the Oracle ADF Code Corner,  which is a blog-style column that provides hints, tips and coding examples for ADF developers. The content on this page ranges from easy to complex and often contains advanced programming concepts. The content is inspired by questions asked on the Oracle JDeveloper customer forum on OTN. The ADF Code Corner has many articles that will inspire your imagination and possibly solve your coding problem.How about the Oracle ADF Architecture Square link? The Oracle ADF Architecture Square focuses on architectural issues and developer guidelines for writing ADF software solutions. The goal is to give ADF developers an understanding of the necessary decisions for building a successful ADF application, to offer potential architectural blueprints to choose from when putting the ADF application together, and to provide potential ADF best practices to take back to your development team. The Oracle ADF Mobile link gives information on developing mobile applications for iOS and Android based applications. There are links to ADF Mobile Overview, ADF Mobile demos and ADF Mobile courses.The Sample ADF Applications link lists sample applications and other resources where you can find code samples for ADF. These are complete ADF applications that can be downloaded into JDeveloper and give you insight into coding an application.There are many more links found under the "Learn More" tab that can equip the developer with the knowledge they need to develop their applications. There are links to overview papers, technical resources, related topics and available training. The information you need IS just a click away.

    Read the article

  • Keeping your options open in a cloud solution

    - by BuckWoody
    In on-premises solutions we have the full range of options open for a given computing solution – but we don’t always take advantage of them, for multiple reasons. Data goes in a Relational Database Management System, files go on a share, and e-mail goes to the Exchange server. Over time, vendors (including ourselves) add in functionality to one product that allow non-standard use of the platform. For example, SQL Server (and Oracle, and others) allow large binary storage in or through the system – something not originally intended for an RDBMS to handle. There are certainly times when this makes sense, of course, but often these platform hammers turn every problem into a nail. It can make us “lazy” in our design – we sometimes don’t take the time to learn another architecture because the one we’ve spent so much time with can handle what we want to do. But there’s a distinct danger here. In nature, when a population shares too many of the same traits, it can cause a complete collapse if a situation exploits a weakness shared by that population. The same is true with not using the righttool for the job in a computing environment. Your company or organization depends on your knowledge as a professional to select the best mix of supportable, flexible, cost-effective technologies to solve their problems, whether you’re in an architect role or not.  So take some time today to learn something new. The way I do this is to select a given problem, and try to solve it with a technology I’m not familiar with. For instance – create a Purchase Order system in Excel, then in Hadoop or MongoDB, or even in flat-files using PowerShell as an interface. No, I’m not suggesting any of these architectures are the proper way to solve the PO problem, but taking something concrete that you know well and applying that meta-knowledge to another platform will assist you in exercising the “little grey cells” and help you and your organization understand what is open to you. And of course you can do all of this on-premises – but my recommendation is to check out a cloud platform (my suggestion would of course be Windows Azure :) ) and try it there. Most providers (including Microsoft) provide free time to do that.

    Read the article

  • MVC: Nasty __o not declared

    - by xamlnotes
    I ran into this little error with MVC where a bunch of errors showed up about  __o  not declared. This was driving me nuts. Then I ran across this link that solved it. http://stackoverflow.com/questions/750902/how-do-i-get-rid-of-o-is-not-declared So, the solution is to put this into the top of the page like VS does for your site.master. <%-- The following line works around an ASP.NET compiler warning --%> <%: ""%> But what about other pages? Lets say you have a view that’s using your site master and that view is throwing this error. Just add the items into the content section where the error occurs like so: <asp:Content ID="indexContent" ContentPlaceHolderID="MainContent" runat="server"> <%-- The following line works around an ASP.NET compiler warning --%> <%: ""%>   Then add the rest of your code. That seems to fix it and its pretty simple too.

    Read the article

  • X server not starting up after new kernel compilation

    - by tech_learner
    I have compiled the Kernel on my 64-bit Debian XPS Studio 1340 Dell system. srikanth@debian:~ - 05:40:52 PM - $ uname -a Linux debian 2.6.32-5-amd64 #1 SMP Thu Mar 22 17:26:33 UTC 2012 x86_64 GNU/Linux Kernel version that I have used and compiled from kernel.org is 2.6.35.13 I have nvidia installed on old kernel. I got the old config and I used the same config to compile the new kernel. Everything went well and I got two debian packages ( image and headers ) which I have installed on my system. When I select the new kernel on the boot menu and I go into it, the X server is not starting up possibly because I have to "rebuild" ( not sure how to do that ) according to this link: http://www.linuxquestions.org/questions/slackware-14/x-server-not-starting-after-kernel-compilation-605265/ Can you suggest how to do the rebuild on nvidia module so that I can start x ( without seeing any blank screen or error saying nvidia module is missing ) ? PS: The link that I have used to compile the kernel is https://help.ubuntu.com/community/Kernel/Compile#Alternate_Build_Method:_The_Old-Fashioned_Debian_Way

    Read the article

  • xubuntu 12.04 screen regularly stops refreshing, refreshing resumes after un-/re-maximizing a window

    - by user68477
    My screen frequently stops completely refreshing. I can make it resume refreshing by un-maximizing/re-maximize a window or by switching workspace (the un-/re-maximizing works every time. Switching workspace sometimes has to be done a couple of times). The immediate impression is that the system is frozen: there is apparently no reaction to anything I do but interestingly window title bar will change, if I switch application with (i.e alt+tab or browse through folders) I saw an identical issue in ubuntu 10.04, though a lot less frequent, I never saw this in ubuntu 12.04 (which I have been using the last 4-5 months). After switching to Xubuntu I'm seeing this again and more frequently. The specific reason I'm not sure this is a bug: I installed gnome-control-center which dragged in tons of packages. This was while trying to fix dual-screen setup. I believe the issue surfaced after this. I later meticulously removed every package from this batch (purge) in the hope that every setting would also be removed. But the issue has persisted. Another issue happened at the same time, it may be totally unrelated but it feels as if it is the same basic issue: the screen resolution of the greeter became less than the expected 1680x1050 and often after login there's just a blank and totally unresponsive wallpaper without panel so I have to force reboot. When the login is successful it's very clear that it works hard to determine the correct resolution which is achieved after a few blink to black screens. My questions: 1) Is this a settings issue or a bug? 2) How do I begin to research the issue - could I perhaps some way reset xubuntu/xfce to default. 3) If this is a bug where would be the most appropriate place report this? System: Thinkpad T500 ati radeon HD 3650 $ fglrxinfo display: :0.0 screen: 0 OpenGL vendor string: Advanced Micro Devices, Inc. OpenGL renderer string: ATI Mobility Radeon HD 3650 OpenGL version string: 3.3.11627 Compatibility Profile Context $ uname -a Linux srvname 3.2.0-32-generic #51-Ubuntu SMP Wed Sep 26 21:33:09 UTC 2012 x86_64 x86_64 x86_64 GNU/Linux xfce 4.10 Compiz 0.9.7.8

    Read the article

  • SMTP POP3 & PST. Acronyms from Hades.

    - by mikef
    A busy SysAdmin will occasionally have reason to curse SMTP. It is, certainly, one of the strangest events in the history of IT that such a deeply flawed system, designed originally purely for campus use, should have reached its current dominant position. The explanation was that it was the first open-standard email system, so SMTP/POP3 became the internet standard. We are, in consequence, dogged with a system with security weaknesses so extreme that messages are sent in plain text and you have no real assurance as to who the message came from anyway (SMTP-AUTH hasn't really caught on). Even without the security issues, the use of SMTP in an office environment provides a management nightmare to all commercial users responsible for complying with all regulations that control the conduct of business: such as tracking, retaining, and recording company documents. SMTP mail developed from various Unix-based systems designed for campus use that took the mail analogy so literally that mail messages were actually delivered to the users, using a 'store and forward' mechanism. This meant that, from the start, the end user had to store, manage and delete messages. This is a problem that has passed through all the releases of MS Outlook: It has to be able to manage mail locally in the dreaded PST file. As a stand-alone system, Outlook is flawed by its neglect of any means of automatic backup. Previous Outlook PST files actually blew up without warning when they reached the 2 Gig limit and became corrupted and inaccessible, leading to a thriving industry of 3rd party tools to clear up the mess. Microsoft Exchange is, of course, a server-based system. Emails are less likely to be lost in such a system if it is properly run. However, there is nothing to stop users from using local PSTs as well. There is the additional temptation to load emails into mobile devices, or USB keys for off-line working. The result is that the System Administrator is faced by a complex hybrid system where backups have to be taken from Servers, and PCs scattered around the network, where duplication of emails causes storage issues, and document retention policies become impossible to manage. If one adds to that the complexity of mobile phone email readers and mail synchronization, the problem is daunting. It is hardly surprising that the mood darkens when SysAdmins meet and discuss PST Hell. If you were promoted to the task of tormenting the souls of the damned in Hades, what aspects of the management of Outlook would you find most useful for your task? I'd love to hear from you. Cheers, Michael

    Read the article

  • Wired Connection Problem

    - by Dave
    After upgrading to 12.04 my interent connection no longer works. More precisely it is really, really, slow, and occasionally will connect, but do so only for a few moments and then disappear again. I am on a Lenovo Workstation e20. Output of ifconfig: eth0 Link encap:Ethernet HWaddr 70:f3:95:00:64:3e inet addr:192.168.1.20 Bcast:192.168.1.255 Mask:255.255.255.0 inet6 addr: fe80::72f3:95ff:fe00:643e/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:7398 errors:0 dropped:74 overruns:0 frame:0 TX packets:6684 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:5407828 (5.4 MB) TX bytes:854343 (854.3 KB) Interrupt:20 Memory:fb120000-fb140000 lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:1587 errors:0 dropped:0 overruns:0 frame:0 TX packets:1587 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:152089 (152.0 KB) TX bytes:152089 (152.0 KB) I am really at a loss for what to do. I am relatively new to Ubuntu, searched the other user questions and couldn't figure this out.

    Read the article

  • “It Isn’t Easy At All; Otherwise, Everyone Would Be Doing It”

    - by Kathryn Perry
    A few months ago, JP Saunders (pictured left), who leads the go-to-market initiatives for the Oracle CX Service offering, kicked off a series of articles about modern customer service. He contends that to take care of customers?and the people that support those customers?companies need to make it easy to deliver consistently great experiences. But it’s not easy; it’s an art. The six posts in The Art of Easy series will help you better understand some of the customer service challenges you face and how to avoid common pitfalls. We pulled them all together here in one post for continuity and easy access. Saunders introduces the series with The Art of Easy: Make It Easy To Deliver Great Customer Service Experiences (Part 1). The Art of Easy: Offer Self Service With the Emphasis on Service (Part 2) by David Fulton (pictured left): David Fulton, Director of Product Management, Oracle Service Cloud, shares five tenets of customer self service that move an organization closer to becoming a modern customer service business. Easy Decisions For Complex Problems (Part 3) by Heike Lorenz (pictured right): Heike Lorenz, Director of Global Product Marketing, Policy Automation, writes about automating service policies to ensure that the correct decisions are being applied to the right people. The goal is to nurture the trusted relationships with customers during complex decision-making processes. Moving at the Speed of Easy (Part 4) by Chris Ulmand (pictured left): Chris Omland, Director of Product Management, Oracle Service Cloud, addresses the need for speed to keep up with customers’ expectations. His advice—start with a platform that enables agile innovation, respects a company’s unique needs, and has proven reliability to protect customer relationships. Knowledge Makes It Easy For Everyone (Part 5) by Nav Chakravarti (pictured rig: Vice President Nav Chakravarti, Oracle Service Cloud, talks about managing the knowledge that customers need and want. He coaches readers on delivering answers to customers’ questions easily, in context, with relevance, reliably, and accurately. Making Easy, Both Effective and Efficient (Part 6) by Melinda Uhland (pictured left): Melinda Uhland, Oracle CX Product Management teaches us that happy agents produce happy customers. A Modern Customer Service organization is one that invests in its agents and empowers them with tools to make them efficient and effective, which, in turn, improves customer results.

    Read the article

  • Is learning C# as a first language a mistake?

    - by JuniorDeveloper1208
    I know there are similar questions on here, which I've read, but I recently read this post by Joel Spolsky: How can I teach a bright person, with no programming experience, how to program? And it got me thinking about my way of learning and whether it might actually be harmful in the long run. I've dabbled with various languages but C# is my first serious one, I've read "Head First C#" and created a few projects. But after reading the post above I've found it a bit disheartening that I may be going about it all wrong, obviously I respect Joel's opinion which is what has thrown me a bit. I've started reading "Code" as recommended in the reading list and I'm finding it pretty hard going, although enjoyable. I feel like it's taken the shine off of my "noobish hacking about" in Visual Studio. So now I'm unsure as to what path I should take? Should I take a step back and follow Joel's advice and start reading? I guess my main aim is just to become a good programmer, like everyone else, but I don't want to be going into bad practice by learning a .NET language when someone who's opinion I respect thinks that it is harmful. Thoughts?

    Read the article

  • linux ftp server with virtual users

    - by kjertil
    i know there are already similar questions for this matter but the answers doesn't really make much sense to anyone who is not really technically comfortable in Linux. I've already tried articles like these for example: http://howto.gumph.org/content/setup-virtual-users-and-directories-in-vsftpd/ with the result of accidently breaking the whole system. The problem is that, while there are several technical possibilities to set up virtual users with a FTP server, it is not as easy as managing for instance a Filezilla server on Windows. I've seen some Web based GUI's but most of them seems to be out of date. The different flavours of Linux and the large amount of different popular FTP servers also seems to make the matter more complicated. I guess my question is, is there a way, to set up virtual FTP users on Linux without the hastle of having to manually edit PAM, MYSQL and config files?

    Read the article

  • SQL Azure Book

    - by ScottKlein
    One of the hotest technology topics of the day is Azure. Being a SQL guy, I am all over this technology, especially SQL Azure. So much so that Herve Roggero and I are currently writing a book for APress on SQL Azure. This book will be out in September and will include deep and thorough coverage of SQL Azure, best practices, and how-to's. We are excited about this book and the technology. However, we'd like to hear from you. As we go around evangelizing SQL Azure at user groups, code camps, and SQL Saturday's, we see the range of "heard of it" to "experimenting with it". Very few are actually doing something with Azure. I'd like to know what your concerns/questions are regarding Azure.  More specifically, what functionality do you think is cool as well as what is lacking, and what would be your list of "must have's" to do something with Azure. We hear a lot regarding security concerns, lack of backup/restore, etc. Is there more? Herve and I will be posting frequently as we write to let you know what we find, what is cool, and what you can look forward to. I'm heading to Tech-Ed in June and hopefully will come back with some great things to tell you!

    Read the article

  • SQL Azure Book

    - by ScottKlein
    One of the hotest technology topics of the day is Azure. Being a SQL guy, I am all over this technology, especially SQL Azure. So much so that Herve Roggero and I are currently writing a book for APress on SQL Azure. This book will be out in September and will include deep and thorough coverage of SQL Azure, best practices, and how-to's. We are excited about this book and the technology. However, we'd like to hear from you. As we go around evangelizing SQL Azure at user groups, code camps, and SQL Saturday's, we see the range of "heard of it" to "experimenting with it". Very few are actually doing something with Azure. I'd like to know what your concerns/questions are regarding Azure.  More specifically, what functionality do you think is cool as well as what is lacking, and what would be your list of "must have's" to do something with Azure. We hear a lot regarding security concerns, lack of backup/restore, etc. Is there more? Herve and I will be posting frequently as we write to let you know what we find, what is cool, and what you can look forward to. I'm heading to Tech-Ed in June and hopefully will come back with some great things to tell you!

    Read the article

  • System reserved

    - by arun sidharth
    I have a HP laptop. I upgraded to windows 7 ultimate from home basic. Now I'm trying to upgrade to Windows 8 but when I do I get a message saying not enough system partition. So I opened the disk manager and increased the size of system reserved partition and it was of no use I still got the same error. Then I unfortunately deleted the 100MB system reserved partition by right clicking it and clicking format in the disk manager! Now I am not able to boot any CD's from the startup including OS and recovery CD. Whenever I press esc it always goes to the login screen and it doesn't say anything about the boot from CD option. Now I could not even use my recovery CD. I have 3 questions: Is it necesseary to create a system reserved partition if so how to create it? How to use my recovery cd How to install win 8

    Read the article

  • Is there a NVIDIA driver (for a 7-series card) that will actually work for 12.10?

    - by DS13
    I see many similar topics on this, but I've tried all their suggestions, and nothing has worked. ISSUE: I do a clean install of Ubuntu 12.10. Boots fine with the “nouveau” graphics driver – graphics are very slow and choppy. The three other driver options in Ubuntu (official NVIDIA drivers), all result in a variation of the black screen on boot up. There will be NO access to a command line/GUI in anyway what-so-ever (tried every option recommended out there, but the system is unusable at this stage). I can only reinstall, and try different drivers…and I only ever get one shot at it. QUESTIONS: Does anyone know of a NVIDIA driver that will actually work with a Nvidia GeForce 7350LE? Or a 7-series card in general? This is my second computer, and I’m just trying to get a working install of Ubuntu on it. I don’t want to put much money into it, as I have seen Ubuntu run great on much older/less capable machines. I’ve got a decent Intel processor (2.3Ghz), 2GB of RAM, 320GB hard drive, 32-bit architecture, and there is no other O/S installed. It appears as if the graphics card is holding me back. Should I just buy a cheap graphics card (non-NVIDIA) to put in as a replacement? TRIED SO FAR: -all drivers available in Ubuntu *all fail -manual install of some different NVIDIA drivers *all fail -also tried installing the generic kernel, Nvidia driver doesn't work in 12.10 *no difference -every method suggested to at least get a command line after switching to a NVIDIA driver *all fail

    Read the article

  • Safari, IIS and optional Client Certificates

    - by Philipp
    I've a ASP.Net Webapp running on IIS7.5. The Webserver is configured to accept Client Certifcates. Unfortunately Visitors with Safari Browser are unable to view the Page. Same Problem as described under the following link: http://www.mnxsolutions.com/apache/safari-providing-an-ssl-error-client-certificate-rejected%E2%80%9D-when-other-browsers-work.html Does anyone knows how to solve this? I'd really appreciate your help. edit: Seems to be the same problem: http://superuser.com/questions/231695/iis7-5-ssl-question-safari-users-get-a-prompt-of-certificate-to-select

    Read the article

  • Can I remove a RAR file's (known) password without recompressing the archive?

    - by Abluescarab
    Long title. Anyway, I haven't been able to find an answer to this question. I know the password to the RAR file, I locked it myself, but now all I want to do is remove the password because it's too much of a pain in the butt to type it in every time. Is there a way to do this in WinRAR or an equivalent program? The only thing I knew to do was to extract it, then create a new RAR without the password. It's not a life-or-death issue, but it would be nice to know. Thanks for your time! EDIT: I just saw a bunch of related questions that appear to ask the same thing. The only solution I saw was using a DOS command to yadda yadda yadda. Here it is: How to remove password protection from compressed files Is there an easier way? Thanks again!

    Read the article

  • Project Jigsaw: On the next train

    - by Mark Reinhold
    I recently proposed to defer Project Jigsaw from Java 8 to Java 9. Feedback on the proposal was about evenly divided as to whether Java 8 should be delayed for Jigsaw, Jigsaw should be deferred to Java 9, or some other, usually less-realistic, option should be taken. The ultimate decision rested, of course, with the Java SE 8 (JSR 337) Expert Group. After due consideration, a strong majority of the EG agreed to my proposal. In light of this decision we can still make progress in Java 8 toward the convergence of the higher-end Java ME Platforms with Java SE. I previously suggested that we consider defining a small number of Profiles which would allow compact configurations of the SE Platform to be built and deployed. JEP 161 lays out a specific initial proposal for such Profiles. There is also much useful work to be done in Java 8 toward the fully-modular platform in Java 9. Alan Bateman has submitted JEP 162, which proposes some changes in Java 8 to smooth the eventual transition to modules, to provide new tools to help developers prepare for modularity, and to deprecate and then, in Java 9, actually remove certain API elements that are a significant impediment to modularization. Thanks to everyone who responded to the proposal with comments and questions. As I wrote initially, deferring Jigsaw to a Java 9 release in 2015 is by no means a pleasant decision. It does, however, still appear to be the best available option, and it is now the plan of record.

    Read the article

  • ideal memory configuration 4 bank, ddr3, AM3+ FX - 1 vs 2 vs 4 dimms?

    - by TardisGuy
    Ok, so ive been looking around, trying to learn and understand the way that ram works. Ive gotten one answer that said "The addressing is best for 2 sticks, and when you use 4; it slows down" Another answer said something like: Theres bank/channel interleave that makes the memory read like one stick Also I read something about the memory density also being a factor. I dug further and found out that theres a higher speed limit on my board for 2 sticks vs 4, so now im trying to put an image in my head of how and why, and... pfft. Can anyone explain, or recommend a resource that would answer these questions?

    Read the article

  • Is (Ubuntu) Linux file copying algorithm better than Windows 7?

    - by Sarath
    Windows Copying is a real mess ever since Windows Vista. Even Microsoft claims they've improved the performance, from a user perspective, it's not quite visible. Even with single file the copying window appears too much time for 'Calculating' and then finishing the copy(Even after 100% completion some times the dialog remains active). At the same time, I was backing up some files in Ubuntu Linux. I felt it's really fast. Might be a feeling caused by faster UI updates. I read an informative post from Jeff Atwood few years back on Windows File Copying. but what my specific questions are Is (Ubuntu) Linux file performance is better than Windows-7? Are both algorithms, Windows and Linux is making use of multiple threads and pipelining mechanism to improve the speed? If yes, which one is better?

    Read the article

  • Photos being copied all over the place

    - by plua
    We have a rather popular website with plenty of photos. Our whole business depends on our content - and the photos are important in this. We invest a lot of time, effort, and money into taking these pictures. On our website we have clear copyright notices, we have the website name and logo in the photos, and we have a Photo Licensing page which states the prices of licensing our photos. Despite all this, our photos are copied by personal and commercial websites alike. We really want to do something about this. We do not want them to take out the photos and leave it at that. We want them to pay for the usage, as we clearly state on our website. Now a few questions come to mind: Can we legally force them to pay right away? Or are we obligated to first write a "Cease and Desist" letter? Photos are used on websites throughout the world. Are there any worldwide rules for this? Has anybody experience with doing these things outside of their home country? Should we hire a lawyer in any country? Or could a local lawyer contact oversees companies directly?

    Read the article

  • error: no such partition after 11.10 upgrade to 12.04

    - by Alan King
    -I recently upgraded my 11.10 install to 12.04 LTS and got the above error message upon reboot after a GNU GRUB version ubuntu3 display showing Ubuntu 3.2.0-23-generic pae and other kernels or memory tests to choose from. The upgrade had to be done by CD because the Update Manager did not show the 12.04 upgrade option. After selecting the default install option of upgrading 11.10 to 12.04, I was presented with a screen saying that I had not specified a swap partition. Upon selection the 'back' key, I was taken to a partition page which listed two current partitions (only Ubuntu 11.10 had been installed - no Windoz): an ext4 partition plus a small 1.8GB partition. I double clicked the small partition and selected it as the swap partition even though I wondered at the time why this even came up. I can see the two user folders under home from the file manager screen while runnning 12.04 from the CD but if I try to access either one an error message is displayed saying I do not have permission while I get a loading message in the lower right corner of the window that does not go away. I have two questions: Can I access the user folders prior to recovery via the Terminal? If so, how? How do I fix the GRUB issue?

    Read the article

  • Simple mail server setup for a user-based website.

    - by Absolute0
    I am working on a simple website that has user registration. All I need to do is confirm email addresses by emailing a confirmation email and maybe send out periodic emails regarding various information. I am also considering having a [email protected] email for user questions and comments. The application is written in django and I am running it on ArchLinux on a VPS. I have no knowledge of mail servers. I tried setting up postfix and qmail but had no luck. They are complicated and require a lot of configuration to get them working properly. My use case is super simple and doesn't not require anything fancy. Is there any simple mailing setup that will enable me to quickly get emailing working with my django app with minimal work?

    Read the article

  • How do I simplify a 2D game grid for level management while keeping its by-pixel features?

    - by Eric Thoma
    (I cross-posted this from StackOverflow as this seems to be a more appropriate forum. I've looked around a little here and I did not find an answer, so I hope this is not a recurring question.) This is a question dealing with 2D world design. I am playing around by creating a 2D bird's eye view shooter game, and I am looking to make the game sleek and advanced. I hope to be able to write physics so projectiles have momentum and knock-down properties. I am immediately running into the problem of world design. I need a way to have level files that store everything there is about a game. This is easiest by just having a grid of objects. But there are thin-walls and other objects that don't seem to fit into a traditional cell of a grid. I want to be able to fit all these together so I can streamline level design; so I don't have to put in the exact pixel-specific start and end of a wall. There doesn't seem to be an obvious translation from level file to game without forcing myself into a pacman-life scenario, meaning a scenario where the game feels boxy and discrete. There is a contrast between the smoothly (relatively) moving characters and finite jumps in a grid. I would appreciate an answer that would describe implementation options or point me to resources that do. I would also appreciate references to sites that teach game design. The language I am using is Java (although I would love to use C or C++, but I can never find convenient resources in those languages). Thank you for any answers. Please leave any questions in the space below; I will be able to answer them later tonight (28th Nov).

    Read the article

  • What are the essential considerations for setting up systems in a location with unreliable power?

    - by dunxd
    I deal with a lot of remote offices located in parts of the world where the local grid power supply is unreliable. Power can go off anytime with no warning, with outages ranging from minutes to days Power fluctuation is wild, with spikes and brown outs Currently the offices will have some or all of the following: A generator, with an inverter, or some sort of manual switch A big UPS or battery array connecting a number of devices Several smaller APC UPS with computers attached Low cost Voltage Regulators sometimes connected between mains and UPS or device. I know that each of these things needs to be appropriately rated for the equipment to which it is connected (although I am not sure how to calculate the correct rating). The offices will generally have the following equipment (in varying quantities): some sort of internet connection device (VSAT router, ADSL modem, WiMax router) Cisco ASA 5505 firewall a bunch of PCs printers one server I don't seek to replace the advice of an electrician, but in some of these locations they only answer the questions you ask them, so I need to make sure I have enough understanding of the essentials to protect equipment from damage, and possibly get through some power cuts.

    Read the article

  • Minimum percentage of free physical memory that Linux require for optimal performance

    - by csoto
    Recently, we have been getting questions about this percentage of free physical memory that OS require for optimal performance, mainly applicable to physical compute nodes. Under normal conditions you may see that at the nodes without any application running the OS take (for example) between 24 and 25 GB of memory. The Linux system reports the free memory in a different way, and most of those 25gbs (of the example) are available for user processes. IE: Mem: 99191652k total, 23785732k used, 75405920k free, 173320k buffers The MOS Doc Id. 233753.1 - "Analyzing Data Provided by '/proc/meminfo'" - explains it (section 4 - "Final Remarks"): Free Memory and Used Memory Estimating the resource usage, especially the memory consumption of processes is by far more complicated than it looks like at a first glance. The philosophy is an unused resource is a wasted resource.The kernel therefore will use as much RAM as it can to cache information from your local and remote filesystems/disks. This builds up over time as reads and writes are done on the system trying to keep the data stored in RAM as relevant as possible to the processes that have been running on your system. If there is free RAM available, more caching will be performed and thus more memory 'consumed'. However this doesn't really count as resource usage, since this cached memory is available in case some other process needs it. The cache is reclaimed, not at the time of process exit (you might start up another process soon that needs the same data), but upon demand. That said, focusing more specifically on the percentage question, apart from this memory that OS takes, how much should be the minimum free memory that must be available every node so that they operate normally? The answer is: As a rule of thumb 80% memory utilization is a good threshold, anything bigger than that should be investigated and remedied.

    Read the article

< Previous Page | 520 521 522 523 524 525 526 527 528 529 530 531  | Next Page >