Search Results

Search found 28517 results on 1141 pages for 'sql und pl sql'.

Page 404/1141 | < Previous Page | 400 401 402 403 404 405 406 407 408 409 410 411  | Next Page >

  • July, the 31 Days of SQL Server DMO’s – Day 21 (sys.dm_db_partition_stats)

    - by Tamarick Hill
    The sys.dm_db_partition_stats DMV returns page count and row count information for each table or index within your database. Lets have a quick look at this DMV so we can review some of the results. **NOTE: I am going to create an ‘ObjectName’ column in our result set so that we can more easily identify tables. SELECT object_name(object_id) ObjectName, * FROM sys.dm_db_partition_stats As stated above, the first column in our result set is an Object name based on the object_id column of this result set. The partition_id column refers to the partition_id of the index in question. Each index will have at least 1 unique partition_id and will have more depending on if the object has been partitioned. The index_id column relates back to the sys.indexes table and uniquely identifies an index on a given object. A value of 0 (zero) in this column would indicate the object is a HEAP and a value of 1 (one) would signify the Clustered Index. Next is the partition_number which would signify the number of the partition for a particular object_id. Since none of my tables in my result set have been partitioned, they all display 1 for the partition_number. Next we have the in_row_data_page_count which tells us the number of data pages used to store in-row data for a given index. The in_row_used_page_count is the number of pages used to store and manage the in-row data. If we look at the first row in the result set, we will see we have 700 for this column and 680 for the previous. This means that just to manage the data (not store it) is requiring 20 pages. The next column in_row_reserved_page_count is how many pages have been reserved, regardless if they are being used or not. The next 2 columns are used for storing LOB (Large Object) data which could be text, image, varchar(max), or varbinary(max) columns. The next two columns, row_overflow, represent pages used for data that exceed the 8,060 byte row size limit for the in-row data pages. The next columns used_page_count and reserved_page_count represent the sum of the in_row, lob, and row_overflow columns discussed earlier. Lastly is a row_count column which displays the number of rows that are in a particular index. This DMV is a very powerful resource for identifying page and row count information. By knowing the page counts for indexes within your database, you are able to easily calculate the size of indexes. For more information on this DMV, please see the below Books Online link: http://msdn.microsoft.com/en-us/library/ms187737.aspx Follow me on Twitter @PrimeTimeDBA

    Read the article

  • HTML Markup in einem APEX Tree - ganz einfach per Plugin!

    - by carstenczarski
    Die APEX Tree Region kennt sicherlich jeder APEX-Entwickler. Und vielfach besteht der Bedarf, das Aussehen des APEX Tree mit Hilfe von HTML Markup zu beeinflussen. Leider ist es seit APEX 4.0 nicht mehr möglich, eigenes HTML-Markup in einen APEX-Tree aufzunehmen - aus Sicherheitsgründen (Schutz vor Cross-Site-Scripting) werden alle HTML Sonderzeichen maskiert. Wenn kein XSS-Risiko besteht (die vom Tree dargestellten Inhalte basieren nicht auf Benutzereingaben und werden komplett vom Entwickler bestimmt), kann dies mit wenigen Zeilen JavaScript und jQuery-Code erreicht werden. Damit es noch einfacher wird,  haben wir die Funktionalität für Sie in einem APEX-Plugin gekapselt. Und so funktioniert es: APEX Plugin "HTML Markup for APEX Tree Region" herunterladenhttp://apex-plugin.com/oracle-apex-plugins/dynamic-action-plugin/html-markup-for-apex-tree_174.html APEX Plugin in die Anwendung importieren APEX Tree Region erzeugen und eigene Ersetzungen für HTML-Sonderzeichen verwenden, also bspw."[" für "<", "]" für ">" und "§" für "&". Eine neue dynamische Aktion erzeugen, die beim Laden der Seite ausgeführt wird und mit Hilfe des Plugins die Ersetzungen im Tree durch die "richtigen" HTML-Sonderzeichen ersetzt. Fertig. Wie das Plugin wirkt, können Sie sich auf einer Demo-Seite ansehen.

    Read the article

  • Using OpenQuery

    - by Derek Dieter
    The OPENQUERY command is used to initiate an ad-hoc distributed query using a linked-server. It is initiated by specifying OPENQUERY as the table name in the from clause. Essentially, it opens a linked server, then executes a query as if executing from that server. While executing queries directly and receiving data directly in this [...]

    Read the article

  • The Windows Azure Software Development Kit (SDK) and the Windows Azure Training Kit (WATK)

    - by BuckWoody
    Windows Azure is a platform that allows you to write software, run software, or use software that we've already written. We provide lots of resources to help you do that - many can be found right here in this blog series. There are two primary resources you can use, and it's important to understand what they are and what they do. The Windows Azure Software Development Kit (SDK) Actually, this isn't one resource. We have SDK's for multiple development environments, such as Visual Studio and also Eclipse, along with SDK's for iOS, Android and other environments. Windows Azure is a "back end", so almost any technology or front end system can use it to solve a problem. The SDK's are primarily for development. In the case of Visual Studio, you'll get a runtime environment for Windows Azure which allows you to develop, test and even run code all locally - you do not have to be connected to Windows Azure at all, until you're ready to deploy. You'll also get a few samples and codeblocks, along with all of the libraries you need to code with Windows Azure in .NET, PHP, Ruby, Java and more. The SDK is updated frequently, so check this location to find the latest for your environment and language - just click the bar that corresponds to what you want: http://www.windowsazure.com/en-us/develop/downloads/ The Windows Azure Training Kit (WATK) Whether you're writing code, using Windows Azure Virtual Machines (VM's) or working with Hadoop, you can use the WATK to get examples, code, PowerShell scripts, PowerPoint decks, training videos and much more. This should be your second download after the SDK. This is all of the training you need to get started, and even beyond. The WATK is updated frequently - and you can find the latest one here: http://www.windowsazure.com/en-us/develop/net/other-resources/training-kit/     There are many other resources - again, check the http://windowsazure.com site, the community newsletter (which introduces the latest features), and my blog for more.

    Read the article

  • Big label generator

    - by jamiet
    Sometimes I write blog posts mainly so that I can find stuff when I need it later. This is such a blog post. Of late I have been writing lots of deployment scripts and I am fan of putting big labels into deployment scripts (which, these days, reside in SSDT) so one can easily see what’s going on as they execute. Here’s such an example from my current project: which results in this being displayed when the script is run: In case you care….PM_EDW is the name of one of our databases. I’m almost embarrassed to admit that I spent about half an hour crafting that and a few others for my current project because a colleague has just alerted me to a website that would have done it for me, and given me lots of options for how to present it too: http://www.patorjk.com/software/taag/#p=testall&f=Banner3&t=PM__EDW Very useful indeed. Nice one! And yes, I’m sure there are a myriad of sites that do the same thing - I’m a latecomer, ok? @Jamiet

    Read the article

  • Merge Join component sorted outputs [SSIS]

    - by jamiet
    One question that I have been asked a few times of late in regard to performance tuning SSIS data flows is this: Why isn’t the Merge Join output sorted (i.e.IsSorted=True)? This is a fair question. After all both of the Merge Join inputs are sorted, hence why wouldn’t the output be sorted as well? Well here’s a little secret, the Merge Join output IS sorted! There’s a caveat though – it is only under certain circumstances and SSIS itself doesn’t do a good job of informing you of it. Let’s take a look at an example. Here we have a dataflow that consumes data from the [AdventureWorks2008].[Sales].[SalesOrderHeader] & [AdventureWorks2008].[Sales].[SalesOrderDetail] tables then joins them using a Merge Join component: Let’s take a look inside the editor of the Merge Join: We are joining on the [SalesOrderId] field (which is what the two inputs just happen to be sorted upon). We are also putting [SalesOrderHeader].[SalesOrderId] into the output. Believe it or not the output from this Merge Join component is sorted (i.e. has IsSorted=True) but unfortunately the Merge Join component does not have an Advanced Editor hence it is hidden away from us. There are a couple of ways to prove to you that is the case; I could open up the package XML inside the .dtsx file and show you the metadata but there is an easier way than that – I can attach a Sort component to the output. Take a look: Notice that the Sort component is attempting to sort on the [SalesOrderId] column. This gives us the following warning: Validation warning. DFT Get raw data: {992B7C9A-35AD-47B9-A0B0-637F7DDF93EB}: The data is already sorted as specified so the transform can be removed. The warning proves that the output from the Merge Join is sorted! It must be noted that the Merge Join output will only have IsSorted=True if at least one of the join columns is included in the output. So there you go, the Merge Join component can indeed produce a sorted output and that’s very useful in order to avoid unnecessary expensive Sort operations downstream. Hope this is useful to someone out there! @Jamiet  P.S. Thank you to Bob Bojanic on the SSIS product team who pointed this out to me!

    Read the article

  • Why would a SQL query JOIN on the same table twice with the same condition?

    - by Scott Leis
    I'm working on changes to a SQL Server v8 database developed by someone else, and have found something that seems (based on my limited SQL knowledge) strange and pointless. One of the views has a query that does a LEFT OUTER JOIN twice on the same table with the same condition. Is there any reason for doing this? The query is below. See the second- and third-last lines that both join the "te_SDE_Survey" table on the "SDE_ID" field. Also note these lines set two different aliases for the table, and both aliases are used in the SELECT part of the query. SELECT vs.SLMS_Code, vs.Retail_Date, vs.TagNo, vs.Rego, vs.Model, vs.Company, vs.AccountType, viqdp.SDE_ID, bd.Debit_Date, isu.Survey_Date, CASE WHEN isu.Q6 IS NOT NULL THEN isu.Q6 ELSE CASE WHEN returned_surveys.survey_date IS NULL THEN CASE WHEN (viqdp.expiryDate < getdate() AND cs.sup1 IS NULL AND cs.sup2 IS NULL AND cs.sup3 IS NULL AND cs.sup5 IS NULL AND cs.sup8 IS NULL AND cs.sup9 IS NULL) THEN 'E' WHEN (viqdp.expiryDate < getdate() AND cs.sup1 = 'F' AND cs.sup2 = 'F' AND cs.sup3 = 'F' AND cs.sup5 = 'F' AND cs.sup8 = 'F' AND cs.sup9 = 'F') THEN 'E' WHEN cs.sup1 = 'T' THEN 'S' WHEN cs.sup2 = 'T' AND (cs.sup8 = 'F' AND cs.sup9 = 'F') THEN 'D' WHEN cs.sup3 = 'T' AND (cs.sup8 = 'F' AND cs.sup9 = 'F') THEN 'D' WHEN cs.sup5 = 'T' AND (cs.sup8 = 'F' AND cs.sup9 = 'F') THEN 'D' WHEN cs.sup8 = 'T' AND (cs.sup2 = 'F' AND cs.sup3 = 'F' AND cs.sup5 = 'F') THEN 'E' WHEN cs.sup9 = 'T' AND (cs.sup2 = 'F' AND cs.sup3 = 'F' AND cs.sup5 = 'F') THEN 'E' WHEN (cs.sup8 = 'T' OR cs.sup9 = 'T') AND (cs.sup2 = 'T' OR cs.sup3 = 'T' OR cs.sup5 = 'T') THEN 'S' END WHEN (tey.survey_expire_method = 'pre2008') THEN CASE WHEN (datediff(month, viqdp.generate_date, returned_surveys.survey_date) 1) THEN 'E' END WHEN (tey.survey_expire_method = 'expiryDateColumn') THEN CASE WHEN (returned_surveys.survey_date viqdp.expiryDate) THEN 'E' END END END AS score_or_exclusion_status, CASE WHEN (bd.explanation IS NULL) THEN '' ELSE bd.explanation END AS explanation, tey.te_Year FROM dbo.te_Vehicle_Sale vs INNER JOIN dbo.te_Year tey ON vs.Retail_Date = tey.Start_Date AND vs.Retail_Date <= tey.End_Date LEFT OUTER JOIN dbo.Bad_Data bd ON vs.TagNo = bd.TagNo LEFT OUTER JOIN dbo.te_Vehicle_SDESurvey viqdp ON vs.TagNo = viqdp.TagNo LEFT OUTER JOIN dbo.te_SDE_Survey isu ON viqdp.SDE_ID = isu.SDE_ID LEFT OUTER JOIN dbo.te_SDE_Survey returned_surveys ON viqdp.SDE_ID = returned_surveys.SDE_ID LEFT OUTER JOIN dbo.te_SDE_Contact_Suppression cs ON viqdp.SDE_ID = cs.SDE_ID

    Read the article

  • Only a few places left for the SQL Social evening on 16th March

    - by simonsabin
    We've got over 50 people registered for the SQLSocial event on 16th March with Itzik Ben-Gan, Greg Low, Davide Mauri and Bill Vaughn I need to finalise numbers on early next week so if you want to come along please register asap, otherwise I can't promise that we'll have space for you. To register use he form on herehttp://sqlsocial.com/events.aspx. I look forward to hearing from you.

    Read the article

  • The PASS Elections Review Committee Needs Your Feedback

    - by andyleonard
    Introduction PASS has had an ERC (Elections Review Committee) forum running for a few months now. There's been surprisingly little feedback, though lots of reads. Here's what it looks like tonight: That's 1,662 views and 37 replies by my count. Not very many replies... Jump In! Now's the time to let PASS know what you think about the current elections process. The ERC members are good people who are trying to make things better. If you have something to add - as simple as "love it!" or "hate it!"...(read more)

    Read the article

  • Database Connectivity Test with UDL File

    - by Ben Griswold
    I bounced around between projects a lot last week.  What each project had in common was the need to validate at least one SQL connection.  Whether you have SQL tools like SSMS installed or not, this is a very easy task if you are aware of the UDL (Universal Data Link) files.  Create a new file and name it anything as long as it has the .udl extension. Open the file, choose a provider: Click Next >> or navigate to the Connection Tab to provide connection information.  Once you provide server and login credentials, the database list will populate.  At this point, you know the connection is valid. but go ahead and click the Test Connection button anyway. On the final tab, you can provide extra connection information like Application Name which can come in handy.  The All tab is beneficial if you want to build a valid connection string to include in your own applications.  If you save the file and then open in Notepad, you’ll find that said connection string: Provider=SQLOLEDB.1;Integrated Security=SSPI;Persist Security Info=False;Initial Catalog=master;Data Source=(local);Application Name=TestApp I hope this tip helps save you some time.  How do you test if you don’t have SSMS installed?

    Read the article

  • Automating SQL Execution Plan analysis

    - by jchang
    Last year, I made my tool for automating execution plan analysis available on www.qdpma.com The original version could parse execution plans from sys.dm_exec_query_stats or dm_exec_cached_plans and generate a cross-reference of which execution plans employed each index. The DMV sys.dm_db_index_usage_stats shows how often each index is used, but not where, that is, which particular stored procedure or My latest version can now also 1) use the DMV sys.dm_exec_procedure_stats, 2) it can also get the...(read more)

    Read the article

  • Learn Who Started that Trace with the Default Trace

    - by Jonathan Kehayias
    This is not Extended Event related but it came from a question on Twitter about how to tell who and from what machine a server side trace was created, and there is no way to explain this in 140 characters so here’s a blog post.  This information is tracked in the Default Trace and can be found by querying for EventClass 175 which is the Audit Server Alter Trace Event trace_event_id from sys.trace_events. select trace_event_id , name from sys . trace_events where name like '%trace%' To query...(read more)

    Read the article

  • TU Berlin wurde mit Oracle Spatial Excellence Award ausgezeichnet

    - by britta wolf
    An der TU Berlin befasst sich das Institut für Geodesie und Geoinformationstechnik von Prof. Thomas Kolbe seit Jahren mit der Verwaltung und Analyse von raumbezogenen Daten in Oracle Spatial. Im Rahmen der diesjährigen Oracle Spatial User Conference in Washington wurden er und sein Team für die herausragenden Ergebnisse der Forschungsarbeiten im Bereich der 3-dimensionalen Stadt- und Geländemodelle ausgezeichnet. Insbesondere die Entwicklung der 3D City Database (3DCityDB), die die Erstellung von virtuellen 3D Stadtmodellen auf Basis von Oracle ermöglicht, sowie das Engagement in der Verbreitung und Weiterentwicklung des CityGML Standards, wurden gewürdigt. Die 3DCityDB mit ihrem Datenbankschema und den zugehörigen Tools steht als Open Source zur Verfügung und wird bereits international in zahlreichen Projekten eingesetzt. Interessiert? Zum Thema Oracle Spatial findet unter dem Titel "Wie kommen die Daten auf die Karte? Anwendungsbeispiele zur Integration von Geodaten" am 11. September um 11:00h ein einführender Webcast statt. Nähere Details dazu gibt es hier.  (Danke an unseren Kollegen Hans Viehmann (aus dem hohen Norden) für diesen Beitrag!)

    Read the article

  • Survey: How much data do you work with?

    - by James Luetkehoelter
    Andy isn't the only one that can ask a survey question. This is something I really curious about because many of the answers or recommendations or rants in blogs are not universably applicable to every database - small databases must sometimes be treated differently, and uber databases are just a pain (and fun at the same time). So, how would you classify most of the databases you work with: 1) Up to 50GB 2) 50-500GB 3) 500GB - 2TB 4) DEAR GOD THAT"S TOO MUCH INFORMATION! Share this post: email it!...(read more)

    Read the article

  • Subsonic 3, SimpleRepository, SQL Server: How to find rows with a null field?

    - by desautelsj
    How ca I use Subsonic's Find<T> method to search for rows with a field containing the "null" value. For the sake of the discussion, let's assume I have a c# class called "Visit" which contains a nullable DateTime field called "SynchronizedOn" and also let's assume that the Subsonic migration has created the corresponding "Visits" table and the "SynchronizedOn" field. If I was to write the SQL query myself, I would write something like: SELECT * FROM Visits WHERE SynchronizedOn IS NULL When I use the following code: var visits = myRepository.Find<Visit>(x => x.SynchronizedOn == null); Subsonic turns it into the following SQL query: SELECT * FROM Visits WHERE SynchronizedOn == null which never returns any rows. I tried the following code but it throws an error: visits = repository.Find<Visit>(x => x.SynchronizedOn.HasValue); I was able to use the following syntax: var query = from v in repository.All<Visit>() where v.SynchronizedOn == null orderby v.CreatedOn select v; visits = query.ToList<Visit>(); but it's not as nice an short as using the Find<T> method. Anyone knows how I can specify the "SynchronizedOn IS NULL" condition in the Find<T> method?

    Read the article

  • Opitz Consulting wins the Oracle SOA Partner Community Award

    - by Jürgen Kress
    Thanks for the nice post! Ein wichtiger Preis für die SOA-Community: Der Oracle EMEA SOA Community Award Im Rahmen der Oracle Open World verlieh Oracle den „Oracle EMEA SOA Community Award" im Bereich "Outstanding Contribution“ im Jahr 2010 bereits zum dritten Mal in Folge an OPITZ CONSULTING: Award als erster SOA Specialized Partner in Europa In 2010 errangen die SOA-Spezialisten von OPITZ CONSULTING den begehrten SOA Partner Community Award. Das Oracle SOA-Team um Jürgen Kress honoriert hiermit das Erreichen der ersten Oracle SOA Spezialisierung in Deutschland, die Community-Arbeit, das Durchführen von SOA-Trainings (auch für anderen Oracle Partner) und das allgemeine Wachstum des OPITZ CONSULTING SOA-Bereichs. Verbindung von EA, BPM und SOA gewürdigt Im Jahr 2009 wurde ein OPITZ CONSULTING Direktor Strategie & Innovation geehrt: Dirk Stähler (Bild li.) erhielt den Award für sein Engagement im Aufbau und der Weiterentwicklung von BPM- und SOA-Themen. Ein besonderer Grund für die Verleihung an ihn waren seine Arbeiten zur effektiven Verbindung von Enterprise Architecture, Business Process Management und SOA. Award für SOA-Community-Arbeit und -Publikationen OPITZ CONSULTING Direktor Strategie & Innovation und Oracle ACE Director, Torsten Winterberg (Bild re.), holte diesen Preis im Jahr 2008 für OPITZ CONSULTING nach Deutschland. Er wurde für sein außerordentliches Engagement in der Etablierung von serviceorientierten Architekturen gewürdigt. Dazu gehören beispielsweise der Aufbau einer Special Interest Group SOA, Roundtable-Initiativen sowie umfangreiche Veröffentlichungen zu SOA.   For more information on the SOA Partner Community please feel free to register at www.oracle.com/goto/emea/soa (OPN account required) Blog Twitter LinkedIn Mix Forum Wiki Website Technorati Tags: SOA Community,Opitz,Opitz Consulting,Torsten Winterberginterberg,oracle,opn,Jürgen Kress

    Read the article

  • Looking Back at PASS Summit 2013 - Location

    - by RickHeiges
    Now that it has been a few weeks since the Summit, I wanted to look back at the location "experiment". Convention Center - It seemed to work well for the conference. There were quite a few areas in the area where you could sit down and get some work down or have a discussion. For the larger welcome reception the first night, I really liked the different areas. If you wanted to enjoy the Quiz Bowl, the ballroom area was set up nicely with big screens so that everyone could see and hear. The area right...(read more)

    Read the article

  • Using Hadooop (HDInsight) with Microsoft - Two (OK, Three) Options

    - by BuckWoody
    Microsoft has many tools for “Big Data”. In fact, you need many tools – there’s no product called “Big Data Solution” in a shrink-wrapped box – if you find one, you probably shouldn’t buy it. It’s tempting to want a single tool that handles everything in a problem domain, but with large, complex data, that isn’t a reality. You’ll mix and match several systems, open and closed source, to solve a given problem. But there are tools that help with handling data at large, complex scales. Normally the best way to do this is to break up the data into parts, and then put the calculation engines for that chunk of data right on the node where the data is stored. These systems are in a family called “Distributed File and Compute”. Microsoft has a couple of these, including the High Performance Computing edition of Windows Server. Recently we partnered with Hortonworks to bring the Apache Foundation’s release of Hadoop to Windows. And as it turns out, there are actually two (technically three) ways you can use it. (There’s a more detailed set of information here: http://www.microsoft.com/sqlserver/en/us/solutions-technologies/business-intelligence/big-data.aspx, I’ll cover the options at a general level below)  First Option: Windows Azure HDInsight Service  Your first option is that you can simply log on to a Hadoop control node and begin to run Pig or Hive statements against data that you have stored in Windows Azure. There’s nothing to set up (although you can configure things where needed), and you can send the commands, get the output of the job(s), and stop using the service when you are done – and repeat the process later if you wish. (There are also connectors to run jobs from Microsoft Excel, but that’s another post)   This option is useful when you have a periodic burst of work for a Hadoop workload, or the data collection has been happening into Windows Azure storage anyway. That might be from a web application, the logs from a web application, telemetrics (remote sensor input), and other modes of constant collection.   You can read more about this option here:  http://blogs.msdn.com/b/windowsazure/archive/2012/10/24/getting-started-with-windows-azure-hdinsight-service.aspx Second Option: Microsoft HDInsight Server Your second option is to use the Hadoop Distribution for on-premises Windows called Microsoft HDInsight Server. You set up the Name Node(s), Job Tracker(s), and Data Node(s), among other components, and you have control over the entire ecostructure.   This option is useful if you want to  have complete control over the system, leave it running all the time, or you have a huge quantity of data that you have to bulk-load constantly – something that isn’t going to be practical with a network transfer or disk-mailing scheme. You can read more about this option here: http://www.microsoft.com/sqlserver/en/us/solutions-technologies/business-intelligence/big-data.aspx Third Option (unsupported): Installation on Windows Azure Virtual Machines  Although unsupported, you could simply use a Windows Azure Virtual Machine (we support both Windows and Linux servers) and install Hadoop yourself – it’s open-source, so there’s nothing preventing you from doing that.   Aside from being unsupported, there are other issues you’ll run into with this approach – primarily involving performance and the amount of configuration you’ll need to do to access the data nodes properly. But for a single-node installation (where all components run on one system) such as learning, demos, training and the like, this isn’t a bad option. Did I mention that’s unsupported? :) You can learn more about Windows Azure Virtual Machines here: http://www.windowsazure.com/en-us/home/scenarios/virtual-machines/ And more about Hadoop and the installation/configuration (on Linux) here: http://en.wikipedia.org/wiki/Apache_Hadoop And more about the HDInsight installation here: http://www.microsoft.com/web/gallery/install.aspx?appid=HDINSIGHT-PREVIEW Choosing the right option Since you have two or three routes you can go, the best thing to do is evaluate the need you have, and place the workload where it makes the most sense.  My suggestion is to install the HDInsight Server locally on a test system, and play around with it. Read up on the best ways to use Hadoop for a given workload, understand the parts, write a little Pig and Hive, and get your feet wet. Then sign up for a test account on HDInsight Service, and see how that leverages what you know. If you're a true tinkerer, go ahead and try the VM route as well. Oh - there’s another great reference on the Windows Azure HDInsight that just came out, here: http://blogs.msdn.com/b/brunoterkaly/archive/2012/11/16/hadoop-on-azure-introduction.aspx  

    Read the article

  • T-SQL Tuesday #13: Clarifying Requirements

    - by Alexander Kuznetsov
    When we transform initial ideas into clear requirements for databases, we typically have to make the following choices: Frequent maintenance vs doing it once. As we are clarifying the requirements, we need to determine whether we want to concinue spending considerable time maintaining the system, or if we want to finish it up and move on to other tasks. Race car maintenance vs installing electric wiring is my favorite analogy for this kind of choice. In some cases we need to sqeeze every last bit...(read more)

    Read the article

  • Merge Join component sorted outputs [SSIS]

    - by jamiet
    One question that I have been asked a few times of late in regard to performance tuning SSIS data flows is this: Why isn’t the Merge Join output sorted (i.e.IsSorted=True)? This is a fair question. After all both of the Merge Join inputs are sorted, hence why wouldn’t the output be sorted as well? Well here’s a little secret, the Merge Join output IS sorted! There’s a caveat though – it is only under certain circumstances and SSIS itself doesn’t do a good job of informing you of it. Let’s take a look at an example. Here we have a dataflow that consumes data from the [AdventureWorks2008].[Sales].[SalesOrderHeader] & [AdventureWorks2008].[Sales].[SalesOrderDetail] tables then joins them using a Merge Join component: Let’s take a look inside the editor of the Merge Join: We are joining on the [SalesOrderId] field (which is what the two inputs just happen to be sorted upon). We are also putting [SalesOrderHeader].[SalesOrderId] into the output. Believe it or not the output from this Merge Join component is sorted (i.e. has IsSorted=True) but unfortunately the Merge Join component does not have an Advanced Editor hence it is hidden away from us. There are a couple of ways to prove to you that is the case; I could open up the package XML inside the .dtsx file and show you the metadata but there is an easier way than that – I can attach a Sort component to the output. Take a look: Notice that the Sort component is attempting to sort on the [SalesOrderId] column. This gives us the following warning: Validation warning. DFT Get raw data: {992B7C9A-35AD-47B9-A0B0-637F7DDF93EB}: The data is already sorted as specified so the transform can be removed. The warning proves that the output from the Merge Join is sorted! It must be noted that the Merge Join output will only have IsSorted=True if at least one of the join columns is included in the output. So there you go, the Merge Join component can indeed produce a sorted output and that’s very useful in order to avoid unnecessary expensive Sort operations downstream. Hope this is useful to someone out there! @Jamiet  P.S. Thank you to Bob Bojanic on the SSIS product team who pointed this out to me!

    Read the article

  • New DMF for SQL Server 2008 sys.dm_fts_parser to parse a string

    Many times we want to split a string into an array and get a list of each word separately. The sys.dm_fts_parser function will help us in these cases. More over, this function will also differentiate the noise words and exact match words. The sys.dm_fts_parser can be also very powerful for debugging purposes. It can help you check how the word breaker and stemmer works for a given input for Full Text Search.

    Read the article

  • I’m now a SQL Server Microsoft Certified Master

    - by simonsabin
    What a day, well what a week really. I found out last week that I passed the Microsoft Certified Master Exam and that I needed to decide to sit the Lab part next. I decided to get it done in a few days of down time I have before going to the MVP summit so scheduled it for today. Five and half hours later I had finished, after a few visits to the toilet. (lesson learnt don’t drink too much during your lab). The lab is marked by humans which I think is a great thing given the various ways of doing...(read more)

    Read the article

  • Oracle 'In Touch' PartnerCast – Neues von der Oracle OpenWorld

    - by Alliances & Channels Redaktion
    Kurz nach den vielen interessanten Ankündigungen auf der Oracle OpenWorld sind alle Oracle EMEA Partner herzlich eingeladen, an dem „In Touch“-Partnerwebcast mit David Callaghan, Senior Vice President EMEA Alliances and Channels, und seinen Studiogästen teilzunehmen. Am Dienstag, 29. Oktober, um 11:30 Uhr MEZ informiert Sie David Callaghan über die neuesten Entwicklungen rund um die Oracle Cloud und Oracle Hardware. Er wird die Key Partner Wins aus dem ersten Quartal vorstellen sowie Ihre Fragen an seine Studiogäste weiterleiten. Live im Studio dabei ist Will O'Brien, Vice President Alliances & Channels, UK & Irland, der über die Highlights der Oracle OpenWorld berichtet und die Relevanz von Storage für das Software Business erläutert. David Callaghan wird Will O’Brien zu seiner neuen Rolle interviewen und ihn nach seinen obersten Prioritäten für FY14 befragen. Markus Reischl, Senior Director und Sales Leader EMEA Strategic Alliances, stellt die Updates in Business Intelligence vor und berichtet darüber, wie diese das gesamte Oracle Portfolio beeinflussen und welche Chancen dies für die Partner bringt. Registrieren Sie sich hier für den Webcast. Über Fragen an die Studiogäste freut sich David Callaghan. Kontaktieren Sie ihn über Twitter mit dem Hashtag #DCpickme oder per E-Mail [email protected]. Mehr Informationen über diesen sowie frühere Webcasts on-demand finden Sie hier auf der neuen Website.

    Read the article

  • Oracle 'In Touch' PartnerCast – Neues von der Oracle OpenWorld

    - by Alliances & Channels Redaktion
    Kurz nach den vielen interessanten Ankündigungen auf der Oracle OpenWorld sind alle Oracle EMEA Partner herzlich eingeladen, an dem „In Touch“-Partnerwebcast mit David Callaghan, Senior Vice President EMEA Alliances and Channels, und seinen Studiogästen teilzunehmen. Am Dienstag, 29. Oktober, um 11:30 Uhr MEZ informiert Sie David Callaghan über die neuesten Entwicklungen rund um die Oracle Cloud und Oracle Hardware. Er wird die Key Partner Wins aus dem ersten Quartal vorstellen sowie Ihre Fragen an seine Studiogäste weiterleiten. Live im Studio dabei ist Will O'Brien, Vice President Alliances & Channels, UK & Irland, der über die Highlights der Oracle OpenWorld berichtet und die Relevanz von Storage für das Software Business erläutert. David Callaghan wird Will O’Brien zu seiner neuen Rolle interviewen und ihn nach seinen obersten Prioritäten für FY14 befragen. Markus Reischl, Senior Director und Sales Leader EMEA Strategic Alliances, stellt die Updates in Business Intelligence vor und berichtet darüber, wie diese das gesamte Oracle Portfolio beeinflussen und welche Chancen dies für die Partner bringt. Registrieren Sie sich hier für den Webcast. Über Fragen an die Studiogäste freut sich David Callaghan. Kontaktieren Sie ihn über Twitter mit dem Hashtag #DCpickme oder per E-Mail [email protected]. Mehr Informationen über diesen sowie frühere Webcasts on-demand finden Sie hier auf der neuen Website.

    Read the article

  • Happy Birthday, SPARC!

    - by A&C Redaktion
    25 Jahre gibt es SPARC in diesem Herbst – da gratulieren Oracle A&C und alle Partner natürlich ganz herzlich! Wir blicken zurück auf ein Vierteljahrhundert Erfolgsgeschichte:Wir befinden uns im Jahr 1987 und klobige graue PCs halten seit einigen Jahren Einzug in Büros und Privathäuser. Ein innovatives Startup-Unternehmen namens Sun Microsystems präsentiert seinen neuen Computer Sun-4, die eigentliche Sensation jedoch ist der Mikroprozessor, den die jungen Leute extra dafür entwickelt hatten: SPARC. Es handelte sich um einen extrem leistungsfähigen RISC-Hauptprozessor, der sowohl in den eigenen Workstations als auch den Servern der Sun-4-Baureihe zum Einsatz kommt. Vor allem in der Unternehmens-IT ermöglicht SPARC in den Folgejahren einen enormen Sprung nach vorn.Die weitere Entwicklung von SPARC, kombiniert mit einem Überblick über andere Meilensteine in der Geschichte der Computerwelt, finden Sie auf der Webseite "Celebrate 25 Years of SPARC Innovation".Wir springen gleich weiter in die Gegenwart, denn auch seit Sun zu Oracle gehört, hat sich so manches getan: Gerade erst hat Oracle die neue Server-Linie Sparc T4 vorgestellt – in Fachkreisen spricht man bereits von der größten Leistungssteigerung in der Geschichte der SPARC-Prozessoren.In den USA wurde das Jubiläum bereits kräftig gefeiert: Hier finden Sie Bilder vom Geburtstagsfest im Museum für Computer-Geschichte in Mountain View, Kalifornien, bei dem auch die SPARC-Entwickler Bill Joy and Andreas von Bechtolsheim zugegen waren und auch im Video SPARC-Event Highlights dreht sich alles um das Jubiläum. In der Oracle Familie gibt es 2012 noch ein weiteres Geburtstagskind: Solaris wird 20, herzlichen Glückwunsch! Das Unix-Betriebssystem, basierend auf SunOS, kam im Jahr 1992 erstmals auf den Markt. Solaris konnte seine gute Stellung seither behaupten und hat nun mit Solaris 11.1 das erste Cloud-Betriebssystem vorgestellt. Dieses überträgt die Zuverlässigkeit, Sicherheit und Skalierbarkeit des bewährten Solaris in die Cloud und bietet eine optimale Plattform für Unternehmensanwendungen.  Lesen Sie hier, was die Fachpresse über die Geburtstagskinder schreibt: ProLinux.de (SPARC) Computerwoche.de (Solaris)SearchDataCenter.de (Solaris)

    Read the article

< Previous Page | 400 401 402 403 404 405 406 407 408 409 410 411  | Next Page >