Search Results

Search found 41254 results on 1651 pages for 'oracle oracle 12c database'.

Page 464/1651 | < Previous Page | 460 461 462 463 464 465 466 467 468 469 470 471  | Next Page >

  • CloudCruiser Chargeback in the cloud

    - by llaszews
    Another company that does chargeback has just been pointed out to me: CloudCruiser There is interesting quote on this company's web site: "Accurate and transparent chargeback is a key requirement in this age of cloud computing. By 2015, we forecast more than 50% of the Global 2000 will charge back most IT costs using service-based pricing, up from less than 10% today. New integrated tools will be needed to implement IT service-based chargeback." - Jay Pultz, Vice President and Distinguished Analyst, Gartner

    Read the article

  • New Nokia SDK 2.0 for Java (beta)

    - by Tori Wieldt
    Nokia recently launched the Asha 305, 306, and 311, which are full touch devices with smartphone-like functionality at a low price. This makes them particularly attractive to consumers in the developing and developed world who may not be able to afford a smartphone but have a strong demand for apps and the smartphone experience. The Asha phones are the latest addition to Nokia's Series 40 platform, all of which support Java ME. The SDK includes new Full Touch API's (e.g. supporting pinch zoom) and Sensor support delivering an enhanced App experience. It also adds improved Maps API support for creating socio-local apps. There are a number of improvements in the tools including the Nokia IDE for Java ME with in-build Device-SDK Manager. Many code examples, training videos, webinars and sample code will help get you started. Porting guides and sample code show you how to port your android app to Java ME. If you don't have access to the hardware you can use Remote Device Access to test on real hardware that's remotely hosted for free. You can also find Light Weight UI Toolkit (LWUIT) support, which can speed development significantly. Both In-App Advertising and In-App Purchase (beta) is supported. Here's a great revenue-making opportunity for developers and a great way of reaching a new app-hungry mass-market audience. Download the new Nokia SDK 2.0 for Java (Beta) and get developing! 

    Read the article

  • Technical Article: Experimenting with Java Timers

    - by Tori Wieldt
    OTN's new tech article is "Experimenting with Java Timers" by T. Lamine Ba. This article studies time—how Java handles timers and the scheduling of tasks. Java timers are utilities that let you execute threads or tasks at a predetermined future time, and these tasks can be repeated according to a set frequency. The article starts with a simple "Hello World" program in a web application that's composed of JavaServer Pages (JSP) and uses the model-control-view (MVC) design pattern. The IDE used in this article is NetBeans IDE 7.1, but you can use any IDE that supports Java. "Experimenting with Java Timers" demonstrates how to get started scheduling jobs with Java. To learn about Swing timers, check out the Java tutorial "How to Use Swing Timers" and additional information in the Java Platform, Standard Edition 7 API Specification for Class Timer. 

    Read the article

  • Best method to implement a filtered search

    - by j0N45
    I would like to ask you, your opinion when it comes to implement a filtered search form. Let's imagine the following case: 1 Big table with lots of columns It might be important to say that this SQL Server You need to implement a form to search data in this table, and in this form you'll have several check boxes that allow you to costumize this search. Now my question here is which one of the following should be the best way to implement the search? Create a stored procedure with a query inside. This stored procedure will check if the parameters are given by the application and in the case they are not given a wildcard will be putted in the query. Create a dynamic query, that is built accordingly to what is given by the application. I am asking this because I know that SQL Server creates an execution plan when the stored procedure is created, in order to optimize its performance, however by creating a dynamic query inside of the stored procedure will we sacrifice the optimization gained by the execution plan? Please tell me what would be the best approach in your oppinion.

    Read the article

  • Learning to program in the modern era?

    - by BBHorus
    At this time, lets say in the modern era, in which order do you organize a programing course for teaching and/or learning, what should be learned first, what should emphasize: Databases Data structures Design patterns Programing paradigms(Procedural, functional, OOP, ...etc ) Operating System Some specific programing language What about English if you are not native speaker or doesn't know English AI Anything else... I ask this because in the university that I went, the programing course was awful it was not focus on what you were going to see out when you work what you were supposed to learn. PS: Again sorry about my English is not my main language. ...Experts and gurus please share

    Read the article

  • Recommendations for SSD for server and database use?

    - by Tony_Henrich
    SSDs are a new technology and they are constantly improving. A lot of the posts here were posted in 2009 when SSDs where less mature and not as fast. What was recommend back then is probably out of date today because of better options. The SSD is used to hold SQL Server databases. Size is probably 128G. The database is used with a CMS and web server so web pages need to get their data and render as fast as possible. Which modern SSD is recommended for such a use? Is there an SSD better than Intel X-25 E/M in terms of performance/cost? (I am also evaluating cost between : RAM + UPS (semi persistent) vs SSD for same amount of gigabytes. No RAID is involved)

    Read the article

  • Hilfreiche Linemode Skripte

    - by Ulrike Schwinn (DBA Community)
    Die mitgelieferten Skripte im Verzeichnis $ORACLE_HOME/rdbms/admin bieten schon seit jeher DBAs und Entwickler eine zusätzliche Unterstützung bei ihrer Arbeit. Sie stehen automatisch nach jeder Installation im Verzeichnis rdbms/admin zur weiteren Verwendung zur Verfügung. Nur wie findet man genau das Skript, das einem die richtige Unterstützung bietet? Eine Dokumentation aller Skripte existiert nicht. Man kann sich an den Namen der Skripte orientieren, da sie mit sprechenden Namen aufgelistet sind und kann die Kurzeinleitung im Skript nachlesen. Das ist mühselig und kostet Zeit. Daher wird im neuen Tipp ein Überblick über wichtige Skripte einschließlich einer Kurzbeschreibung gegeben. Mehr dazu hier

    Read the article

  • Is 10% too much for autogrow on a 4 GB sql server DB?

    - by ntsue
    I am getting the following error: 2011-03-07 21:59:35.73 spid64 Autogrow of file 'MYDB_DATA' in database 'MYDB' was cancelled by user or timed out after 16078 milliseconds. Use ALTER DATABASE to set a smaller FILEGROWTH value for this file or to explicitly set a new file size. I did some research, and I found that for large databases you should set autogrow to a fixed size (MB), and not to a percentage. I feel like this database is not large and I may not be addressing the correct issue by changing this value. Does anyone have any opinions? Thank you! EDIT: I should have specified SQL Server 2008 RC2 running on Windows Server 2008

    Read the article

  • How to manage the images for my Desktop Application

    - by NonExistent
    What's the better way to manage the image files of my app? i've been thinking about the way that i do right now (save the image as a BLOB IN db), and i ask myself if would be better to manage the image as text in my DB, i mean, convert the image to hex(length of 500), then save in the db as text, and when calling it convert it from hex to image, or something like that, but what do you consider as an Experienced Progammer that is the better way? Maybe the question is too broad, but i need to know that, and nobody answers me anywhere...

    Read the article

  • VDC Research Webcast: Engineering Business Value in the IoT with Java 8

    - by tangelucci
    Date: Thursday, June 19, 2014 Time: 9:30 AM PDT, 12:30 PM EDT, 17:30 GMT The growth of the Internet of Things (IoT) opens up new service-driven opportunities, delivering increased efficiencies, better customer value, and improved quality of life. Realizing the full potential of the Internet of Things requires that we change how we view and build devices. These next-generation systems provide the core foundation of the services, rapidly transforming data to information to value. From healthcare to building control systems to vehicle telematic systems, the IoT focuses on how conneted devices can become more intelligent, enhance interoperability with other devices, systems and services, and drive timely decisions while delivering real business return for all. Join this webcast to learn about: Driving both revenue opportunities and operational efficiencies for the IoT value chain Leveraging Java to make devices more secure How Java can help overcome resource gaps around intelligent connected devices Suggestions on how to better manage fragmentation in embedded devices Register here: http://event.on24.com/r.htm?e=793757&s=1&k=4EA8426D0D31C60A2EDB139635FF75AB

    Read the article

  • Stores Still Matter In The World Of E-Commerce

    - by Michael Hylton
    You may think that more and more consumers are moving their purchasing to the Web or mobile device.  However, according The NPD Group, Inc., a leading market research company, 15-20 percent of consumers in 2011 checked out products in stores before buying online, or called “showrooming”, for product categories like stand mixers, electric knives, sewing machines, and some floor cleaners. Other categories like power tools, hairsetters, and robotic vacuums are now beginning to show signs of the “showrooming” trend as well. It is doubly important to present a consistent, personalized, and relevant shopping experience for your customers, no matter whether they interact with you in-person in your store, with your sales agents or call center agents, over the Web, or using a mobile device.  Your goal is to make that experience across touchpoints as seamless as possible.

    Read the article

  • Stores Still Matter In The World Of E-Commerce

    - by Michael Hylton
    You may think that more and more consumers are moving their purchasing to the Web or mobile device.  However, according The NPD Group, Inc., a leading market research company, 15-20 percent of consumers in 2011 checked out products in stores before buying online, or called “showrooming”, for product categories like stand mixers, electric knives, sewing machines, and some floor cleaners. Other categories like power tools, hairsetters, and robotic vacuums are now beginning to show signs of the “showrooming” trend as well. It is doubly important to present a consistent, personalized, and relevant shopping experience for your customers, no matter whether they interact with you in-person in your store, with your sales agents or call center agents, over the Web, or using a mobile device.  Your goal is to make that experience across touchpoints as seamless as possible.

    Read the article

  • How to choose NoSQL database engine?

    - by Poma
    We have a database with following specs: 30k records, 7mb in size 20 inserts/second 1000 updates/second 1000 range selects/second, by secondary index, approx 10 rows each needs at least one secondary index needs some mechanism to expire keys if they are not updated for 75 secs (can be done via programmatic garbage collector but will require additional 'last_update' index and will add some load) consistency is not required durability is not required db should be stored in memory For now we use Redis, but it does not have secondary index and it's keys index:foo:* is too slow. Membase also does not have secondary index (as far as I know). MongoDB and MySQL memory engine have table-level locks. What engine will fit our use case?

    Read the article

  • Lesetipp: Der TDAzlan Exadata Value Guide

    - by Alliances & Channels Redaktion
    Fast 20 Jahre sind Oracle und TDAzlan schon erfolgreiche Partner. Als VAD unterstützt TDAzlan Oracle Partner bei deren Business, zum Beispiel durch Sales- und Marketing-Support, Webcasts, Zertifizierungstrainings oder Enablement Workshops bis hin zur Spezialisierung. Jetzt wurde der Exadata Value Guide von Azlan für Partner und Kunden aktualisiert auf aktuelle 12c und X4 Technologien. Die Broschüre bündelt Fakten und Tipps rund um die Themen Exadata und Engineered Systems. Sie schlüsselt technische Details zu Oracle Software- und Hardware-Komponenten auf und bietet Verkaufsargumente sowie hilfreiche Erläuterungen zum Salesprozess. Außerdem wird das Oracle Authorized Solution Center (OASC) von TDAzlan in München vorgestellt. Dort haben Partner die Möglichkeit, Oracle Gesamtlösungen zu testen und Ihren Kunden zu präsentieren. Den Exadata Value Guide können sie auch auf der Oracle Informations-Webseite von TDAzlan herunterladen.

    Read the article

  • Multitenancy in SQL Azure

    - by cibrax
    If you are building a SaaS application in Windows Azure that relies on SQL Azure, it’s probably that you will need to support multiple tenants at database level. This is short overview of the different approaches you can use for support that scenario, A different database per tenant A new database is created and assigned when a tenant is provisioned. Pros Complete isolation between tenants. All the data for a tenant lives in a database only he can access. Cons It’s not cost effective. SQL Azure databases are not cheap, and the minimum size for a database is 1GB.  You might be paying for storage that you don’t really use. A different connection pool is required per database. Updates must be replicated across all the databases You need multiple backup strategies across all the databases Multiple schemas in a database shared by all the tenants A single database is shared among all the tenants, but every tenant is assigned to a different schema and database user. Pros You only pay for a single database. Data is isolated at database level. If the credentials for one tenant is compromised, the rest of the data for the other tenants is not. Cons You need to replicate all the database objects in every schema, so the number of objects can increase indefinitely. Updates must be replicated across all the schemas. The connection pool for the database must maintain a different connection per tenant (or set of credentials) A different user is required per tenant, which is stored at server level. You have to backup that user independently. Centralizing the database access with store procedures in a database shared by all the tenants A single database is shared among all the tenants, but nobody can read the data directly from the tables. All the data operations are performed through store procedures that centralize the access to the tenant data. The store procedures contain some logic to map the database user to an specific tenant. Pros You only pay for a single database. You only have a set of objects to maintain and backup. Cons There is no real isolation. All the data for the different tenants is shared in the same tables. You can not use traditional ORM like EF code first for consuming the data. A different user is required per tenant, which is stored at server level. You have to backup that user independently. SQL Federations A single database is shared among all the tenants, but a different federation is used per tenant. A federation in few words, it’s a mechanism for horizontal scaling in SQL Azure, which basically uses the idea of logical partitions to distribute data based on certain criteria. Pros You only have a single database with multiple federations. You can use filtering in the connections to pick the right federation, so any ORM could be used to consume the data. Cons There is no real isolation at that database level. The isolation is enforced programmatically with federations.

    Read the article

  • October 2012 Security "Critical Patch Update" (CPU) information and downloads released

    - by user12244672
    The October 2012 security "Critical Patch Update" information and downloads are now available from My Oracle Support (MOS). See http://www.oracle.com/technetwork/topics/security/alerts-086861.html and in particular Document 1475188.1 on My Oracle Support (MOS), http://support.oracle.com, which includes security CVE mappings for Oracle Sun products. For Solaris 11, Doc 1475188.1 points to the relevant SRUs containing the fixes for each issue.  SRU12.4 was released on the CPU date and contains the current cumulative security fixes for the Solaris 11 OS. For Solaris 10, we take a copy of the Recommended Solaris OS patchset containing the relevant security fixes and rename it as the October CPU patchset on MOS.  See link provided from Doc 1475188.1 Doc 1475188.1 also contains references for Firmware, etc., and links to other useful security documentation, including information on Userland/FOSS vulnerabilities and fixes in https://blogs.oracle.com/sunsecurity/

    Read the article

  • PHP - Data Access Layer

    - by scarpacci
    I am currently reviewing a code base and noticed that a majority of the calls (along with DB connections) are just buried inside the PHP scripts. I would have assumed that like other languages they would have developed some sort of data access layer (Like I would do in .Net or Java) for all of the communication to the DB (or implemented MVC, etc). Is this still a common pattern in PHP or is there alternative methodologies/patterns for this technology? I am just trying to understand why the subs would have developed it this way. Any insight/info on how experienced developers design an approach data access in PHP would be very much appreciated.

    Read the article

  • Create .exe and .msi Installers for your JavaFX Apps

    - by mikew_co
    Well I wanted to figure out how the new JavaFX native packaging worked, so with a little work, I have written up my findings. Create a Windows Native Installer and EXE with JavaFX and NetBeans 7.2All the information is in the articles I have linked to at the end of the page. However, I hopefully have pulled together all the key facts in one place. In addition, I tried to document all the problems I ran into in the troubleshooting section. So what is the end result?  With everything installed, building an application in NetBeans creates an EXE installer, an MSI installer, and an EXE file to execute your application. Really slick professional stuff. This is a great addition to the whole Java platform.

    Read the article

  • Why creating a new MDX language instead of extending SQL?

    - by DReispt
    I have a long experience with SQL, but recently began working with datawarehouse and OLAP technologies: building fact and dimension tables, that then are queried using MDX (MultiDimensional eXpressions). The problem is that MDX works with a completely different logic compared to SQL, and it's a whole new learning curve even for someone with a strong SQL background. Yes, MDX allows you to do things that would be hard or almost impossible with plain SQL. But sometimes it's frustrating to be hours around an MDX to do something you know you could achieve in minutes using SQL (ok, you can tell me to RTFM ...). But why go on to the trouble of creating a new completely different language when you could build on SQL, extend it to add the features needed by OLAP applications?

    Read the article

  • There's Not an App for That (Yet)

    - by Mark Hesse
    With an earlier-than-normal departure this morning to avoid the stalemate known as traffic congestion, I suddenly realized what I had failed to grab on my way out the door...  my company ID badge.  Unfortunately, at the time of my epiphany, I was far enough into commuter no-man's land where turning back would completely negate my early departure and increase my overall drive time exponentially.  Not being one to retrace my steps, I decided to press on. Upon arrival at the office and with an hour to go before a security guard would be on duty, I started thinking about the number of times I had forgotten my ID vs. the number of times I had forgotten my phone.  While rare on both accounts, my ID was most likely the missing artifact. I then wondered why there isn't an app for my smartphone that allows me to verify my credentials with my employer and then, provided with a secure token for the day, have the ability to access my building's card entry system.  On many levels, this seems much more secure than an ID card which can be lost, stolen or even forged and then used simply by tailgating into and around buildings at facilities where card scanning can generally be avoided.   As it turns out, another building on the campus has 24 x 7 guard coverage, so I was able to gain access in a relatively short time and secure a temporary ID badge.  Once inside and online, a quick internet search on the subject of smartphone badge access shows that efforts are underway to do exactly what I was thinking needed to be done. Having not spent any time studying about the technology, I discovered that it relies on Near Field Communications (NFC) enabled smartphones (of which, mine does not provide).  The only other option would require modifications to the security infrastructure to support alternative authentication technologies, such as barcode readers, which would be extremely costly to implement. For now, my best option is to put my corporate ID under my car keys... 

    Read the article

  • 12.1.3 Spares Management Enhancements Transfer of Information (TOI)

    - by Oracle_EBS
    Transfer of Information (TOI) presentation is available. It covers the following enhancements made to the EBS Spares Management Product: Restrict Sources with no Shipping Network definition Create Internal Order when Source is Manned Warehouse Display Delivery status in Parts Requirement UI Order Sources by distance when Shipping cost remains same Calculate Parts Shipping Distances using Navteq Data Consider Warehouse Calendar to calculate Parts Arrival Date Create Requisitions in Operating Unit of Destination Inventory Org Uptake of HZ address structure in Parts Requirement UI

    Read the article

  • Tell us what you want in 6.6

    - by Reggie
    Now that we have finished 6.5 it's time to really start gearing up for 6.6.  We have  many great feature ideas but we want to hear from you.  To help with that, we're running a poll on the website where you can vote on the features you would like to see in the next version of the best MySQL ADO.Net Connector on the planet.  You can find the poll at http://dev.mysql.com and scrolling near the bottom of the page.   Thanks for your time and please go vote.  

    Read the article

  • Problems with Ranking when Using Sourcing Rules And ASLs From Blanket Agreements?

    - by LisaO
    Are you using Sourcing Rules and Approved Supplier List with Blanket Purchase Agreements (BPA) and it seems like Ranking is not working correctly? For example:  The Sourcing Rule being used, has effective dates from 01-APR to 31-MAR for 2013, 2014 and 2015. One BPA is defined for Supplier A, which was originally set to Rank 1 with 100% allocation. A new BPA was created for the same item and with same effective dates as the current BPA. The BPA is for a different Supplier. When Generate Sourcing Rules is run it adds the new BPA/Supplier to the Sourcing rule, but its added as Rank 1, with the old rule changed to Rank 2. For complete information refer to  Doc ID 1678447.1 Generate Sourcing Rules And ASLs From Blanket Agreements Ranking not Behaving As Expected. Still have Questions? Access the Procurement Community and, using the 'Start a Discussion' link, post your question.

    Read the article

  • Tweaking Hudson memory usage

    - by rovarghe
    Hudson 3.1 has some performance optimizations that greatly reduces its memory footprint. Prior to this Hudson used to always hold the entire data model (all jobs and all builds) in memory which affected scalability. Some installations configured heap sizes in excess of 1GB to counteract this. Hudson 3.1.x maintains an MRU cache and only loads jobs and builds as they are required. Because of the inability to change existing APIs and be backward compatible with plugins, there were limits to how far we could go with this approach. Memory optimizations almost always come with a related cost, in this case its additional I/O that has to be performed to load data on request. On a small site that has frequent traffic, this is usually not noticeable since the MRU cache will usually hold on to all the data. A large site with infrequent traffic might experience some delays when the first request hits the server after a long gap. If you have a large heap and are able to allocate more memory, the cache settings can be adjusted to take advantage of this and even go back to pre-3.1 behavior. All the cache settings can be passed as options to the JVM container (Tomcat or the default Jetty container) using the -D option. There are two caches, independant of each other, one for Jobs and the other for Builds. For the jobs cache: hudson.jobs.cache.evict_in_seconds ( default=60 ) Seconds from last access (could be because of a servlet request or a background cron thread) a job should be purged from the cache. Set this to 0 to never purge based on time. hudson.jobs.cache.initial_capacity ( default=1024 ) Initial number of jobs the cache can accomodate. Setting this to the number of jobs you typically display on your Hudson landing page or home page will speed up consecutive access to that page. If the default is too large you may consider downsizing and using that memory for the Builds cache instead. hudson.jobs.cache.max_entries ( default=1024) Maximum number of jobs in the cache. The default is large enough for most installations, but if you find I/O activity when always accessing the hudson home page you might consider increasing this, but first verify if the I/O is caused by frequent eviction (see above), rather than by the cache not being large enough. For the builds cache: The builds cache is used to store Build objects as they are read from storage. Typically this happens when a user drills down into the details of a particular Job from the hudson hom epage. The cache is shared among builds for different jobs since in most installations all jobs are not accessed with the same frequency, so a per-job builds cache would be a waste of memory. hudson.job.builds.cache.evict_in_seconds ( default=60 ) Same as the equivalent Job cache, applied to Build. hudson.job.builds.cache.initial_capacity" ( default=512 ) Same as equivalent Job cache setting. Note the smaller initial size. If your site stores a large number of builds and has frequent access to more builds you might consider bumping this up. hudson.job.builds.cache.max_entries ( default=10240 ) The default max is large enough for most installations, the builds cache has bigger sized objects, so be careful about increasing the upper limit on this. See section on monitoring below. Sample usage: java -jar hudson-war-3.1.2-SNAPSHOT.war -Dhudson.jobs.cache.evict_in_seconds=300 \ -Dhudson.job.builds.cache.evict_in_seconds=300 Monitoring cache usage The 'jmap' tool that comes with the JDK can be used to monitor cache performance in an indirect way by looking at the number of Job and Build objects in each cache. Find the PID of the hudson instance and run $ jmap -histo:live <pid | grep 'hudson.model.*Lazy.*Key$' Here's a sample output: num #instances #bytes class name 523: 28 896 hudson.model.RunMap$LazyRunValue$Key 1200: 3 96 hudson.model.LazyTopLevelItem$Key These are the keys to the Jobs (LazyTopLevelItem$Key) and Builds (RunMap$LazyRunValue$Key) in the caches, so counting the number of keys is a good indicator of the number of items in the cache at any given moment. The size in bytes can be ignored, they are just the size of the keys, not the actual sizes of the objects they hold. Those sizes can only be obtained with a profiler. With the output above we can conclude that there are 3 jobs and 28 builds in memory. The 28 builds can all be from 1 job or all 3 jobs. Over time on an idle system, these should get evicted and memory cache should be empty. In practice, because of background cron threads and triggers, jobs rarely fall down to zero. Access of a job or a build by a cron thread resets the eviction timer.

    Read the article

< Previous Page | 460 461 462 463 464 465 466 467 468 469 470 471  | Next Page >