Search Results

Search found 41254 results on 1651 pages for 'oracle oracle 12c database'.

Page 465/1651 | < Previous Page | 461 462 463 464 465 466 467 468 469 470 471 472  | Next Page >

  • Garbage Collection Basics

    - by mikew_co
    Java Is an awesome programming language and platform. One of its better features is automatic garbage collection. Ever wondered how that works? I have written an online web course outlining the basics. Much of what is included has been published before in various white papers and such. However, this is updated for JDK 7 and includes some nice illustrations of the steps involved. Hope you like it. Garbage Collection Basics. A follow-on web course on the G1 garbage collector should follow in a week or so.

    Read the article

  • What is the best way to work with large databases in Java depending on context?

    - by user19000
    We are trying to figure out the best practice for working with very large DBs in Java. What we do is a kind of BI (business Intelligence), i.e analyzing very large DBs, and using them to create intermediate DBs that represent intelligent knowledge of the DBs. We are currently using JDBC, and just preforming queries using a ResultSet. As more and more data is being created, we are wondering whether more appropriate ways exist for parsing and manipulating these large DBs: We need to support 'chunk' manipulation and not an entire DB at once(e.g. limit in JDBC, very poor performance) We do not need to be constantly connected since we are just pulling results and creating new tables of our own. We want to understand JDBC alternatives, with respect to advantages and disadvantages. Whether you think JDBC is the way to go or not, what are the best practices to go by depending on context (e.g. for large DBs queried in chunks)?

    Read the article

  • BIP 11g Dynamic SQL

    - by Tim Dexter
    Back in the 10g release, if you wanted something beyond the standard query for your report extract; you needed to break out your favorite text editor. You gotta love 'vi' and hate emacs, am I right? And get to building a data template, they were/are lovely to write, such fun ... not! Its not fun writing them by hand but, you do get to do some cool stuff around the data extract including dynamic SQL. By that I mean the ability to add content dynamically to your your query at runtime. With 11g, we spoiled you with a visual builder, no more vi or notepad sessions, a friendly drag and drop interface allowing you to build hierarchical data sets, calculated columns, summary columns, etc. You can still create the dynamic SQL statements, its not so well documented right now, in lieu of doc updates here's the skinny. If you check out the 10g process to create dynamic sql in the docs. You need to create a data trigger function where you assign the dynamic sql to a global variable that's matched in your report SQL. In 11g, the process is really the same, BI Publisher just provides a bit more help to define what trigger code needs to be called. You still need to create the function and place it inside a package in the db. Here's a simple plsql package with the 'beforedata' function trigger. Spec create or replace PACKAGE BIREPORTS AS whereCols varchar2(2000); FUNCTION beforeReportTrig return boolean; end BIREPORTS; Body create or replace PACKAGE BODY BIREPORTS AS   FUNCTION beforeReportTrig return boolean AS   BEGIN       whereCols := ' and d.department_id = 100';     RETURN true;   END beforeReportTrig; END BIREPORTS; you'll notice the additional where clause (whereCols - declared as a public variable) is hard coded. I'll cover parameterizing that in my next post. If you can not wait, check the 10g docs for an example. I have my package compiling successfully in the db. Now, onto the BIP data model definition. 1. Create a new data model and go ahead and create your query(s) as you would normally. 2. In the query dialog box, add in the variables you want replaced at runtime using an ampersand rather than a colon e.g. &whereCols.   select     d.DEPARTMENT_NAME, ...  from    "OE"."EMPLOYEES" e,     "OE"."DEPARTMENTS" d  where   d."DEPARTMENT_ID"= e."DEPARTMENT_ID" &whereCols   Note that 'whereCols' matches the global variable name in our package. When you click OK to clear the dialog, you'll be asked for a default value for the variable, just use ' and 1=1' That leading space is important to keep the SQL valid ie required whitespace. This value will be used for the where clause if case its not set by the function code. 3. Now click on the Event Triggers tree node and create a new trigger of the type Before Data. Type in the default package name, in my example, 'BIREPORTS'. Then hit the update button to get BIP to fetch the valid functions.In my case I get to see the following: Select the BEFOREREPORTTRIG function (or your name) and shuttle it across. 4. Save your data model and now test it. For now, you can update the where clause via the plsql package. Next time ... parametrizing the dynamic clause.

    Read the article

  • Webcast: CRM Foundations - Notes, Attachments and Folder Technology

    - by LuciaC
    Webcast: CRM Foundations - Notes, Attachments and Folder Technology Date: November 21, 2013 at 11am ET, 10am CT, 8am PT, 4pm GMT, 9.30pm IST Don't miss this webcast if you want to know how to get the most out of using Notes and learn how to leverage best practices for Folder technology and Attachments.  This session will help users who are struggling with any of these topics understand how to use them better and more efficiently. TOPICS WILL INCLUDE: Set up and use of Notes Notes Security Attachments and their use throughout CRM Folder Technology Any new functionality related to these topics in release 12.2Set up and use of Notes. For more details and how to register see Doc ID 1592459.1 Remember that you can access a full listing of all future webcasts as well as replays from Doc ID 7409661.1.

    Read the article

  • Iterative and Incremental Principle Series 5: Conclusion

    - by llowitz
    Thank you for joining me in the final segment in the Iterative and Incremental series.  During yesterday’s segment, I discussed Iteration Planning, and specifically how I planned my daily exercise (iteration) each morning by assessing multiple factors, while following my overall Implementation plan. As I mentioned in yesterday’s blog, regardless of the type of exercise or how many increment sets I decide to complete each day, I apply the 6 minute interval sets and a timebox approach.  When the 6 minutes are up, I stop the interval, even if I have more to give, saving the extra energy to apply to my next interval set.   Timeboxes are used to manage iterations.  Once the pre-determined iteration duration is reached – whether it is 2 weeks or 6 weeks or somewhere in between-- the iteration is complete.  Iteration group items (requirements) not fully addressed, in relation to the iteration goal, are addressed in the next iteration.  This approach helps eliminate the “rolling deadline” and better allows the project manager to assess the project progress earlier and more frequently than in traditional approaches. Not only do smaller, more frequent milestones allow project managers to better assess potential schedule risks and slips, but process improvement is encouraged.  Even in my simple example, I learned, after a few interval sets, not to sprint uphill!  Now I plan my route more efficiently to ensure that I sprint on a level surface to reduce of the risk of not completing my increment.  Project managers have often told me that they used an iterative and incremental approach long before OUM.   An effective project manager naturally organizes project work consistent with this principle, but a key benefit of OUM is that it formalizes this approach so it happens by design rather than by chance.    I hope this series has encouraged you to think about additional ways you can incorporate the iterative and incremental principle into your daily and project life.  I further hope that you will share your thoughts and experiences with the rest of us.

    Read the article

  • Interesting fact #123423

    - by Tim Dexter
    Question from a customer on an internal mailing list this, succintly answered by RTF Template God, Hok-Min Q: Whats the upper limit for a sum calculation in terms of the largest number BIP can handle? A: Internally, XSL-T processor uses double precession.  Therefore the upper limit and precision will be same as double (IEEE 754 double-precision binary floating-point format, binary64). Approximately 16 significant decimal digits, max is 1.7976931348623157 x 10308 . So, now you know :)

    Read the article

  • T-SQL Tuesday #015:Remaining on Alert

    - by AllenMWhite
    This T-SQL Tuesday is about Automation in SQL Server, and I'll bet you think I'm going to talk about PowerShell. Well, you're partially right. What I'd like to talk about today, though, is setting up alerts to automate some responses to naturally occuring phenomena on your SQL Servers. (Note: I first introduced this technique in an article on Simple Talk here ). Without going back to the original article, I'm talking about a safety valve process for when your transaction logs start to fill at rates...(read more)

    Read the article

  • T-SQL Tuesday #015:Remaining on Alert

    - by AllenMWhite
    This T-SQL Tuesday is about Automation in SQL Server, and I'll bet you think I'm going to talk about PowerShell. Well, you're partially right. What I'd like to talk about today, though, is setting up alerts to automate some responses to naturally occuring phenomena on your SQL Servers. (Note: I first introduced this technique in an article on Simple Talk here ). Without going back to the original article, I'm talking about a safety valve process for when your transaction logs start to fill at rates...(read more)

    Read the article

  • Possibility Program for number of pieces

    - by Brad
    I would like to put a program together to calculate the number of 60' pieces would be needed from a list of shorter pieces. For example, I sell rebar cut to length from our standard length of 60'-0". Now the length the customer requires are as follows: 343 pc @ 12.5' 35 pc @ 13' 10 pc @ 15' 63 pc @ 15.5'....... There are 56 total lengths ranging from 12.5' to 30.58' The idea is to limit the amount of waste from the 60' piece. The input from the user would be: number of differnt lengths Length of piece to cut from count of different lengths The result would be the number of prime pieces needed to fulfill the order. What well-known algorithms exist that could help me solve this problem?

    Read the article

  • Building MySQL with boost on windows

    - by user13177919
    As you've probably heard already MySQL needs boost to build. However, in the good ol' MySQL tradition, the above link does give you only the instructions on how to build it on linux. And completely ignores the fact that there're other OSes too that people develop on. To fill in that gap, I've compiled a small step by step guide on how to do it on windows. Note that I always, as a principle, build out-of-source. The typical setup I have is : bzr clone lp:~mysql/mysql-server/5.7 mysql-trunkcd mysql-trunkmkdir bldcd bldcmake -DWITH_DEBUG=1 -DMYSQL_PROJECT_NAME=mysql-trunk ..devenv /build debug mysql-trunk.sln This has been tested to work on a 32 bit compile using VS2013 on a Windows7 64 bit build. Note that you'll need other things too (bison, eventually openssl etc) that I will assume you already have set up. Steps: Download Boost 1.55.0. It's the *only* version that is known to work currently. Extract boost_1_55_0/ from the zip to c:\boost\boost_1_55_0 Go to Control Panel/System/Environment variables and set WITH_BOOST=C:\boost\boost_1_55_0 in User variables. Make sure you restart your open command line terminal windows after this !  If you're upgrading from non-boost build, remove your bld/ directory and create a new one. run cmake as you'd typically do. You should get: -- Local boost dir C:/boost/boost_1_55_0 -- Local boost zip LOCAL_BOOST_ZIP-NOTFOUND -- BOOST_VERSION_NUMBER is #define BOOST_VERSION 105500 -- BOOST_INCLUDE_DIR C:/boost/boost_1_55_0 Build as normal (devenv /build debug ...). It should work.

    Read the article

  • Need data on disk drive management by OS: getting base I/O unit size, “sync” option, Direct Memory A

    - by Richard T
    Hello All, I want to ensure I have done all I can to configure a system's disks for serious database use. The three areas I know of (any others?) to be concerned about are: I/O size: the database engine and disk's native size should either match, or the database's native I/O size should be a multiple of the disk's native I/O size. Disks that are capable of Direct Memory Access (eg. IDE) should be configured for it. When a disk says it has written data persistently, it must be so! No keeping it in cache and lying about it. I have been looking for information on how to ensure these are so for CENTOS and Ubuntu, but can't seem to find anything at all! I want to be able to check these things and change them if needed. Any and all input appreciated.

    Read the article

  • (My) Sun Ray 3i

    - by user13346636
    Last week, some Sun Ray devices were shown at the LASDEC exhibition. Afterward, they were brought back to the Aoyama Center, but not all of them found a place to be stored. So, two days ago, Iwasaki-san, one of the co-workers I've been close to (and who was at the exhibition), put a Sun Ray 3i (all-in-one with 21.5" screen) on my (shared) desk. Yay! I managed to get a Japanese keyboard, and now I can access my card and cardless sessions from Germany, and the performance is just great, as good as when I work from home in Hamburg. That's the way my deskt looks now,almost as messy as my desk in Hamburg: And my back is very grateful.

    Read the article

  • Universal Work Queue Quick Filter Examples

    - by LuciaC-Oracle
    If you use Universal Work Queue then it's likely that you will want to define and use your own Quick Filters.  Quick Filters allow you to focus on specific work classes based on assigned criteria in a node. This makes it much easier for Agents to view their work grouped in a meaningful way.How to create Universal Work Queue - Quick Filters (Doc ID 803163.1) gives two worked examples to help you understand how to create your own Quick Filters:     Adding a 'Resource Group' filter     Adding an Overdue Amounts filter for use in Collections. We hope you find these examples useful.  Let us know by providing feedback on the document itself or, why not post to the MOS Service Community with your experience and suggestions.

    Read the article

  • Our work name lately transformed to Revenue from Customer Support... Support? [on hold]

    - by Hollis Nieves
    I have been employeed from the same company (mobile phone business) for quite some time today like a Customer Support Represenative. I also have usually completed nicely until recently and loved the task. Into a Person Support/Revenue middleapproximately we've been converted by May and we've to purchase Television service to clients who call-in about their mobile phone service... They need us to become really manipulative with it. We havent had instruction on it. and our supervisors actually dont understand something about any of it possibly but need us to "Purchase! Purchase! Purchase!" ugggh. At I will become at my work any guidance? Revenue 've never be completed by me and I truly worry losing my work due to my performance. Revenue makes me uneasy... Any feedback could be appreciated

    Read the article

  • Resize a new database to predicted maximum size

    - by John Oxley
    Currently I have a SQL Server database which is about 2 Gb. I know over the next year it's going to grow to a maximum of about 10Gb. Hard drive space is not an issue in the slightest. Is there a down side to resizing the datafile to 20Gb now, then defragmenting the hard drive? Should I resize the log file to 1Gb as well? Something ridiculously large so that fragmentation doesn't happen there either. With this question I would like to avoid the datafile becoming fragmented on the disk itself, but I don't want to negatively impact performance.

    Read the article

  • Tutorial on Hudson, JUnit and Ant

    - by Grant Ronald
    Often when discussing ADF we often show the features for developing applications. However, writing applications is only one part.  Building in a team, integrating code, testing it...these are equally important to the success of the project.  If you would like to find out how features in JDeveloper can help you build, maintain, integrate and test your application then check out this tutorial.

    Read the article

  • CRM Webcast: Territory Setup and Manage Matching Attributes

    - by LuciaC
    Subject:  Territory Setup and Manage Matching Attributes Date: July 9, 2013 at 1pm ET, 12pm CT, 11am MT, 10am PT, 6pm, BST (London, GMT+01:00), 10:30 pm IST (Mumbai, GMT+05:30)Territories are used in a number of different EBS CRM applications, including Sales, Field Service and Service Contracts.  If you want to know more about how territories work and how to set them up, join our experts in this webcast.  The webcast will a demonstrate a high level setup for one of the Sales products and examples of how other applications use the Territory Manager. Topics will include: Enabling Matching Attributes Custom Matching Attributes Examples for Account, Leads, Quote, Proposals, Opportunities in the Sales product. Running Concurrent Requests Details & Registration: Doc ID 1544622.1

    Read the article

  • Mission Critical: SQL Server Upgrade

    You're faced with the task of doing a SQL Server Upgrade? Do you know all the steps, and the right order to do them? You do? Even with interruptions and distractions? Maybe, but it is wise to be able to refer to the Mission-Critical Task checklist.

    Read the article

  • Consolidate. Compress. Control. -- Oracle Database 11g Release 2

    While managing exponentially increasing data volumes and meeting demands for high availability, DBAs are also tasked with delivering ever greater service with ever greater efficiency. Find out how Oracle Database 11g Release 2 can help DBAs address these challenges and help companies control and manage their information—while lowering costs.

    Read the article

  • OracleDBConsoleORCL service cannot be started

    - by sasola
    I installed oracle in my own machine from about two years, but I stopped some oracle services from about seven months. Now I tried to start them again and open the oracle db but I couldn't make a connection, the OracleDBConsoleORCL service cannot be started although it was not one of the services that I stopped, it was starting automatically! I tried to make it manually started and start it again but the problem didn't be solved. Could someone help me?

    Read the article

  • Good choice of languages for making a program that manages and organizes business? [closed]

    - by Ronney P
    I've been reading questions and discussions on this website but haven't made an account to start talking or asking anything I had doubts in so please bare with a newbie here. What are specific languages that have are able to make a program that will record, and organize things such as hours, salaries, payments? Also solve business problems, mostly with payments, how much money there will be after interest, taxes and such. Anyway, I've been looking into COBOL, C++, Java, HTML, JavaScript, VB.NET and a couple more. Which ones should I focus on and look into more? I very much appreciate any answers. Thank you.

    Read the article

  • Live HTML Preview

    - by netbeanstips
    When you need to edit an HTML document in NetBeans you may find useful this little plugin that adds a Preview tab to HTML editor window. The plugin works with some small issues in NetBeans 7.4 but I recommend using development builds instead. So install the plugin, restart NetBeans and open any HTML document. Notice there's a Preview button in editor's toolbar (see the red rectangle in the picture below). Now split the editor window by dragging the split button at top right corner. You can also use menu View - Split - Vertically.  Then in the bottom split part toggle Preview button. You will get a live preview of your HTML source code. The preview pane will auto-refresh as you edit the HTML code. There are even some handy tools in Preview toolbar, for example you can resize the preview browser to match the screen dimensions of various device types. I know there is full-blown HTML5 support in NetBeans 7.4. But if you need to edit a single document or when you're running Java-only NetBeans distribution this plugin may come handy... Note: The plugin is built on top of embedded WebKit browser which is based on JavaFX WebView component. So there might be some issue when using the plugin on some flavors of Linux.

    Read the article

  • Learning MySQL Query optimization

    - by recluze
    I've been doing web/desktop/server development for a while and have worked with many databases (mysql mostly). I've come to the point in my career when I need to have someone look at my queries because they're 'kind of slow'. I believe it's now time to start learning query optimization. While I know the basics of index and joins etc., I'm not familiar with how to use, say, the EXPLAIN output to improve performance of my queries. I have not been able to find any online material that starts with the basics and takes me to application. Getting a book is not an option right now so I'm looking for tips about how to proceed with this. I hope this question is general enough not to get closed.

    Read the article

  • Is it good practice to keep 2 related tables (using auto_increment PK) to have the same Max of auto_increment ID when table1 got modified?

    - by Tum
    This question is about good design practice in programming. Let see this example, we have 2 interrelated tables: Table1 textID - text 1 - love.. 2 - men... ... Table2 rID - textID 1 - 1 2 - 2 ... Note: In Table1: textID is auto_increment primary key In Table2: rID is auto_increment primary key & textID is foreign key The relationship is that 1 rID will have 1 and only 1 textID but 1 textID can have a few rID. So, when table1 got modification then table2 should be updated accordingly. Ok, here is a fictitious example. You build a very complicated system. When you modify 1 record in table1, you need to keep track of the related record in table2. To keep track, you can do like this: Option 1: When you modify a record in table1, you will try to modify a related record in table 2. This could be quite hard in term of programming expecially for a very very complicated system. Option 2: instead of modifying a related record in table2, you decided to delete old record in table 2 & insert new one. This is easier for you to program. For example, suppose you are using option2, then when you modify record 1,2,3,....,100 in table1, the table2 will look like this: Table2 rID - textID 101 - 1 102 - 2 ... 200 - 100 This means the Max of auto_increment IDs in table1 is still the same (100) but the Max of auto_increment IDs in table2 already reached 200. what if the user modify many times? if they do then the table2 may run out of records? we can use BigInt but that make the app run slower? Note: If you spend time to program to modify records in table2 when table1 got modified then it will be very hard & thus it will be error prone. But if you just clear the old record & insert new records into table2 then it is much easy to program & thus your program is simpler & less error prone. So, is it good practice to keep 2 related tables (using auto_increment PK) to have the same Max of auto_increment ID when table1 got modified?

    Read the article

< Previous Page | 461 462 463 464 465 466 467 468 469 470 471 472  | Next Page >