Search Results

Search found 15797 results on 632 pages for 'session variables'.

Page 359/632 | < Previous Page | 355 356 357 358 359 360 361 362 363 364 365 366  | Next Page >

  • JavaOne Kicks Off with Sunday Keynotes

    - by Yolande Poirier
    The Java Strategy, Partner, and Technical keynotes will be held on Sunday, September 22, beginning at 4:00 p.m. like last year, to free up time for session slots on Monday and Tuesday. The keynotes will again take place at the historic Masonic Auditorium on Nob Hill. That same evening at 7:00 p.m., attendees are invited to the official JavaOne Welcome Reception at the Taylor Street Café @ the Zone. Sunday will also feature User Group meetings (at Moscone West) and Java University courses (Hilton San Francisco Union Square). On Thursday, the Java Community keynote will start the wrap up of the conference. Register before July 19, 2013 and save US$400. Click here for information on registration packages, including the low-cost Discover pass alternative.

    Read the article

  • Google dévoile enfin Android 3.0 Honeycomb, dans lequel "tout est pensé pour les tablettes"

    Google dévoile enfin Android 3.0 Honeycomb, dans lequel "tout est pensé pour les tablettes" Depuis le temps qu'on en parlait, le voilà ! Google vient de présenter officiellement Honeycomb, la nouvelle mouture d'Android, lors d'un évènement spécial. Cette session était diffusée en streaming live sur le Web, ce qui nous a permis de la suivre en direct, dès son commencement à 19 heures (heure française). Android 3.0 est surtout attendu parce qu'il est adapté aux tablettes tactiles, son principe étant en effet de simplifier à l'extrême l'utilisation des appareils mobiles, en accédant à leurs différentes fonctions le plus facilement et directement possible. Et ça se sent dès l'écran d'accueil, qui...

    Read the article

  • New Survey Findings: Application Intelligence and Connected Devices - How do you Harness the Value

    - by Yolande Poirier
    Oracle and Beecham have recently conducted a market survey on use of Connected Devices for M2M & Internet of Things (IoT) applications and new trends. This first session in our webinar series addresses intelligence in connected devices. Join Peter Utzschneider of Oracle and Robin Duke-Woolley of Beecham Research as they discuss: What are the key business drivers of your connected devices program? To what extent do you expect the intelligence required for M2M & IoT applications to change? Would these changes occur at the network edge, at the data center, or both? What are the impacts of these changes on ISV’s and device manufacturers? What are the opportunities for other M2M & IoT players?

    Read the article

  • Open XML at TechEd 2010

    Open XML was a big part of my first session at TechEd 2010 called, "Office 2010: Developing the Next Wave of Productivity Solutions". The thing that gets the biggest reaction is the Open XML SDK 2.0 "Productivity Tool"-- especially the ability to reflect over an Office document to produce C# code that will produce the target document. Here's the scenario: I have a Word document (Excel spreadsheet, PowerPoint deck) that a user produced manually. I want to be able to produce that same document...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • More Sessions At Central Coast Code Camp, Ruby/Cloud Computing

      Should Your Application Run In The Cloud Im back and sitting in Steve Evans Session, Should Your Application Run In The Cloud.  Hes now explaining how computers, since the stone age,... This site is a resource for asp.net web programming. It has examples by Peter Kellner of techniques for high performance programming...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Product Support Webcast for Existing Customers: Oracle Webcenter Portal 11g User & Administration Tips

    - by John Klinke
    Register for our upcoming Advisor Webcast 'Oracle WebCenter Portal 11g: User & Administration Tips' scheduled for November 12, 2013 at 11:00 am, Eastern Standard Time (8:00 am Pacific Standard Time, 4:00 pm GMT Time, 5:00 pm Europe Time). This 1-hour session is recommended for technical and functional users who use Oracle WebCenter Portal to build company portals using run-time tools.Topics will include:• Whats new in 11.1.1.8 of WebCenter Portal• Terminology Changes• Using the Portal once its built• Setting up Self Registration (Admins)• End User Experience• Development Environment• Patching InformationFor more information and to register for this Advisor Webcast, please see Oracle WebCenter Portal 11g: User & Administration Tips (Doc ID 1585902.1).

    Read the article

  • SSIS Tips & Tricks (Presentation)

    This has been a rather well used presentation title but it does allow a certain degree of flexibility, and we covered a good range of topics in my session at the UK SQL Server User Group in Cambridge last night. Thanks to all who attended. Here is the rather limited slide deck and the all important demo packages for download as promised. For reference, high level topics covered were BIDS Helper Inserts and Updates Transactions Script Debugging Data Flow Checkpoints I’ll update the post with a link to the Live Meeting recording when I get it. Presentation & Demo Packages (194KB) SSIS Tips & Tricks - Darren Green.zip

    Read the article

  • Standard -server to server- and -browser to server- authentication method

    - by jeruki
    I have server with some resources; until now all these resources were requested through a browser by a human user, and the authentication was made with an username/password method, that generates a cookie with a token (to have the session open for some time). Right now the system requires that other servers make GET requests to this resource server but they have to authenticate to get them. We have been using a list of authorized IPs but having two authentication methods makes the code more complex. My questions are: Is there any standard method or pattern to authenticate human users and servers using the same code? If there is not, are the methods I'm using now the right ones or is there a better / more standard way to accomplish what I need? Thanks in advance for any suggestion.

    Read the article

  • Remote GDM login

    - by luri
    Is it possible to remotely (via ssh) make a login to GDM? I'll explain a situation in which I'd need this: I'm on my laptop and connect to my desktop via ssh. I realize there are some updates available. I install them all, but I need to reboot the computer. The problem is that once the desktop has rebooted, there will be no GDM session available, since it's necessary to log in. Can I somehow perform the GDM login remotely from my laptop? Edit: I can't find FreeNX for Maverick, and I haven't been able to configure x11vnc + vncviewer to work. Isn't there an easy way?

    Read the article

  • How do you unit test your javascript.

    - by Erin
    I spend a lot of time working in javascript of late. I have not found a way that seems to work well for testing javascript. This in the past hasn't been a problem for me since most of the websites I worked on had very little javascript in them. I now have a new website that makes extensive use of jQuery I would like to build unit tests for most of the system. My problems are this. Most of the functions make changes to the DOM in some way. Most of the functions request data from the web server as well and require a session on the service to get results back. I would like to run the test from either a command line or a test running harness rather then in a browser. Any help or articles I should be reading would be helpful.

    Read the article

  • What Are the Windows A: and B: Drives Used For?

    - by Jason Fitzpatrick
    The C: drive is the default installation location for Windows, if you have a CD/DVD drive on your machine it’s likely the D: drive, and any additional drives fall in line after that. What about the A: and B: drives? Image by Michael Holley. Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-drive grouping of Q&A web sites. HTG Explains: What is DNS? How To Switch Webmail Providers Without Losing All Your Email How To Force Windows Applications to Use a Specific CPU

    Read the article

  • Oracle Database Appliance Technical Boot Camp

    - by mseika
    Oracle Database Appliance Technical Boot Camp Wednesday 19th September 9.30 – 16.30 This session is designed to give our partners detailed sales and technical information to familiarise themselves with the Oracle Database Appliance. It is split into two sessions, the first aimed at sales and pre-sales technical support, and the second aimed at pre-sales and technical implementation staff. The agenda is as follows: Part 1 Oracle Engineered Systems Introducing the Oracle Database Appliance What is the target market? Competitive positioning Sales Plays Up sell opportunities Resell requirements and process Part 2 Hardware internals Download the appliance software kit Disabling / enabling cores Configuration and setup Oracle 11g R2 overview Backup strategies Please register here.

    Read the article

  • Product Support Webcast for Existing Customers: WebCenter Content 11.1.1.8 Overview and Support Information

    - by John Klinke
    Register for our upcoming Advisor Webcast 'WebCenter Content 11.1.1.8 Overview and Support Information' scheduled for 11am Eastern, November 21, 2013 (10am Central, 9am Mountain, 8am Pacific, 17:00 Europe Time (Paris, GMT+01:00)).This 1-hour session is recommended for technical and functional users who have installed or plan to upgrade to WebCenter Content 11.1.1.8 or would just like more information on the latest release.Topics will include: Overview of new features and enhancements Installation of the new WebCenter Content web UI Upgrading from older WebCenter Content versions Support issues including latest patches Roadmap of proposed additional features Make sure you register and mark this date on your calendar. Register at: https://oracleaw.webex.com/oracleaw/onstage/g.php?d=590991341&t=aOnce the host approves your request, you will receive a confirmation email with instructions for joining the call on November 21st.

    Read the article

  • Openbox overhead is similar to that with gnome-panel

    - by drN
    I just installed openbox via sudo apt-get install openbox. It already has obconf, btw. I noticed that when I logged into my openbox session instead of the one I usually use (gnome-panel and NOT Ubuntu 2D or one of the high overhead environments) and checked via htop, I found that a similar amount of RAM was being occupied (~600 MB or so) with openbox or gnome-panel. What gives? Openbox looks lighter but it certainly isn't any different. Obviously the same daemons etc would run in both environments as they share the same folders. Is gnome-panel as good as openbox then?

    Read the article

  • The Ubuntu Advantage Service right for my non-profit?

    - by Robert
    My small 5 computer office currently runs on Ubuntu. 2 of the desktops run Windows7 in Sun Virtualbox software, and are used for Quickbooks. I am going off to college, and I am looking for a paid tech support solution to replace my IT position. I have an approx $300/mth budget, and I am wiling to discuss higher rates. Everyone in the office is currently comfortable with regular desktop usage, but I am handling all of the software installation and updates. I was hoping to get a total support package for all of their tech related questions, but I cannot find any services which will support linux. Is the Ubuntu Advantage Service something which can take my place? They would mostly need network help, printer help, and an occasional software compatibly troubleshooting session. If this is not a solution, does anyone know of a tech support forum/hotline which would cover all of this? Thank you for reading.

    Read the article

  • Ubuntu 12.04, web cam light flashing

    - by user205877
    Have been running Ubuntu on my Compaq C757 reliably since 9.10. Currently running a fresh install of 12.04 LTS. Somewhere during my regular updates, my webcam started flashing its power light. Light flashes from logon screen all throughout my session. I am not running Google Chrome but chromium and firefox. Webcam works normally whenever I run Cheese webcam booth, light comes on solid and webcam works normally. I close out of Cheese, webcam shuts off, and then the light starts blinking again. My system is dual boot, when I boot into Windows 7, webcam does not exhibit this problem. I suspect issue may be related to new laptop power management settings in kernel based on research on the web of similar problems, but I do not know how to instruct it to treat webcam differently.

    Read the article

  • Slides, Scripts, and Photos from SharePoint Saturday New Orleans 2011

    - by Brian Jackett
    This weekend I presented “Managing SharePoint 2010 Farms with PowerShell” at SharePoint Saturday New Orleans.  This was my first time visiting New Orleans so I was excited for the experience.  A big thanks to everyone who attended my session.  I condensed the material a little but the slides and scripts below have additional material that we couldn’t cover.  Let me know if you have any comments, questions, or feedback.  Thanks. Slides and Scripts     Managing SharePoint 2010 Farms with PowerShell   Photos     <coming soon since the conference is still ongoing>         -Frog Out

    Read the article

  • Webcast - C-level Perspectives on How HR Can Take on a Bigger Role in Strategic Planning

    - by Scott Ewart
    The Economist Intelligence Unit (EIU), on behalf of IBM and Oracle, recently surveyed a number of C-level executives in North America and Western Europe to understand how HR can take on a bigger role in driving growth. The resulting reports highlight the actions senior HR leaders can take to place themselves at the heart of the debate on a company's strategic direction.In this session, IBM and Oracle HCM specialists will review the findings of the EIU research reports and provide guidance on how technology innovation can help to align talent strategies with long term business goals. Participants will gain an understanding of the following: Results of the Economist Intelligence Unit study around "Executive Perceptions of the HR Function" Differences in perspective between CEOs and CFOs Identify how the HR professional can take a bigger role in driving business growth Join us on Thursday, October 25 for a live webcast. Speakers:Gina Wells Global Oracle HCM LeaderIBM Global Business Services Michelle NewellSenior Director, HCM Applications MarketingOracle Register Here For the Webcast on Thursday, October 25.

    Read the article

  • SQL SERVER – Backing Up and Recovering the Tail End of a Transaction Log – Notes from the Field #042

    - by Pinal Dave
    [Notes from Pinal]: The biggest challenge which people face is not taking backup, but the biggest challenge is to restore a backup successfully. I have seen so many different examples where users have failed to restore their database because they made some mistake while they take backup and were not aware of the same. Tail Log backup was such an issue in earlier version of SQL Server but in the latest version of SQL Server, Microsoft team has fixed the confusion with additional information on the backup and restore screen itself. Now they have additional information, there are a few more people confused as they have no clue about this. Previously they did not find this as a issue and now they are finding tail log as a new learning. Linchpin People are database coaches and wellness experts for a data driven world. In this 42nd episode of the Notes from the Fields series database expert Tim Radney (partner at Linchpin People) explains in a very simple words, Backing Up and Recovering the Tail End of a Transaction Log. Many times when restoring a database over an existing database SQL Server will warn you about needing to make a tail end of the log backup. This might be your reminder that you have to choose to overwrite the database or could be your reminder that you are about to write over and lose any transactions since the last transaction log backup. You might be asking yourself “What is the tail end of the transaction log”. The tail end of the transaction log is simply any committed transactions that have occurred since the last transaction log backup. This is a very crucial part of a recovery strategy if you are lucky enough to be able to capture this part of the log. Most organizations have chosen to accept some amount of data loss. You might be shaking your head at this statement however if your organization is taking transaction logs backup every 15 minutes, then your potential risk of data loss is up to 15 minutes. Depending on the extent of the issue causing you to have to perform a restore, you may or may not have access to the transaction log (LDF) to be able to back up those vital transactions. For example, if the storage array or disk that holds your transaction log file becomes corrupt or damaged then you wouldn’t be able to recover the tail end of the log. If you do have access to the physical log file then you can still back up the tail end of the log. In 2013 I presented a session at the PASS Summit called “The Ultimate Tail Log Backup and Restore” and have been invited back this year to present it again. During this session I demonstrate how you can back up the tail end of the log even after the data file becomes corrupt. In my demonstration I set my database offline and then delete the data file (MDF). The database can’t become more corrupt than that. I attempt to bring the database back online to change the state to RECOVERY PENDING and then backup the tail end of the log. I can do this by specifying WITH NO_TRUNCATE. Using NO_TRUNCATE is equivalent to specifying both COPY_ONLY and CONTINUE_AFTER_ERROR. It as its name says, does not try to truncate the log. This is a great demo however how could I achieve backing up the tail end of the log if the failure destroys my entire instance of SQL and all I had was the LDF file? During my demonstration I also demonstrate that I can attach the log file to a database on another instance and then back up the tail end of the log. If I am performing proper backups then my most recent full, differential and log files should be on a server other than the one that crashed. I am able to achieve this task by creating new database with the same name as the failed database. I then set the database offline, delete my data file and overwrite the log with my good log file. I attempt to bring the database back online and then backup the log with NO_TRUNCATE just like in the first example. I encourage each of you to view my blog post and watch the video demonstration on how to perform these tasks. I really hope that none of you ever have to perform this in production, however it is a really good idea to know how to do this just in case. It really isn’t a matter of “IF” you will have to perform a restore of a production system but more of a “WHEN”. Being able to recover the tail end of the log in these sever cases could be the difference of having to notify all your business customers of data loss or not. If you want me to take a look at your server and its settings, or if your server is facing any issue we can Fix Your SQL Server. Note: Tim has also written an excellent book on SQL Backup and Recovery, a must have for everyone. Reference: Pinal Dave (http://blog.sqlauthority.com)Filed under: Notes from the Field, PostADay, SQL, SQL Authority, SQL Performance, SQL Query, SQL Server, SQL Tips and Tricks, T SQL

    Read the article

  • MEF to the rescue in Task-it

    One of the frameworks that I chose to leverage in the development of my Silverlight 4 Task-It application is MEF, the Managed Extensibility Framework. Actually, that decision came while I was at a Silverlight Deep Dive session on the Microsoft campus in Redmond, WA, and came at the recommendation of John Papa (Microsoft Silverlight Evangelist) and Glenn Block (the MEF Maniac). If you are not familiar with MEF, its primary function is to provide a framework for extensibility in Silverlight applications, however, what I found is that MEF's capabilities go beyond just this. As a matter of fact, Microsoft announced at MIX this week that its new Silverlight Analytics Framework would use MEF for composition. You can read about Telerik's support for this new framework here: Telerik is First to Announce Support for Microsoft Silverlight Analytics Framework MEF used to be just one of many frameworks out there that could ...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Join Companies in Web and Telecoms by Adopting MySQL Cluster

    - by Antoinette O'Sullivan
    Join Web and Telecom companies who have adopted MySQL Cluster to facilitate application in the following areas: Web: High volume OLTP eCommerce User profile management Session management and caching Content management On-line gaming Telecoms: Subscriber databases (HLR/HSS) Service deliver platforms VAS: VoIP, IPTV and VoD Mobile content delivery Mobile payments LTE access To come up to speed on MySQL Cluster, take the 3-day MySQL Cluster training course. Events already on the schedule include:  Location  Date  Delivery Language  Berlin, Germany  16 December 2013  German  Munich, Germany  2 December 2013  German  Budapest, Hungary  4 December 2013  Hungarian  Madrid, Spain  9 December 2013  Spanish  Jakarta Barat, Indonesia  27 January 2014  English  Singapore  20 December 2013  English  Bangkok, Thailand  28 January 2014  English  San Francisco, CA, United States  28 May 2014  English  New York, NY, United States  17 December 2013  English For more information about this course or to request an additional event, go to the MySQL Curriculum Page (http://education.oracle.com/mysql).

    Read the article

  • Pagination and duplicate content

    - by jazz090
    I have an archive page that displays the number of articles published. Because there were so many, I ran a pagination script: for 127.0.0.1/archive/2/?p=x&pp=y where p is the page number and pp is number of articles to display per page. The pagination looks like this: Prev 1 2 3 4 ... 12 NEXT with each item linking to p like <a href="?p=x">x</a>. I also have the items per page setter: 25 | 50 | 100 (<a href="?pp=y">y</a>). Now I have a PHP script that fixes pp into a session variable. But I am worried about duplicate content (since incrementing pp values will be inclusive) and also content not getting indexed because its not in the pagination link. so in the example above, pages 5-11 will not be indexed. Any ideas on how to fix this?

    Read the article

  • Community Forum at Openworld - Presentations available

    - by Javier Puerta
    On October 1st we held a new session of the Exadata & Manageability Partner Community in San Francisco. Thanks to all of you who participated in the event and very especially to the partner speakers who share their experiences with the rest of the community: Francisco Bermúdez (Capgemini Spain), Dmitry Krasilov (Nvision, Russia) and Miguel Alves (WeDo Technologies, Portugal)The slide decks used in the presentations are now available for download at the Manageability Partner Community Collaborative Workspace (for community members only - if you get an error message, please register for the Community first).In a few weeks we will be announcing the location for the next Community event in the spring timeframe.

    Read the article

  • Restoring databases to a set drive and directory

    - by okeofs
     Restoring databases to a set drive and directory Introduction Often people say that necessity is the mother of invention. In this case I was faced with the dilemma of having to restore several databases, with multiple ‘ndf’ files, and having to restore them with different physical file names, drives and directories on servers other than the servers from which they originated. As most of us would do, I went to Google to see if I could find some code to achieve this task and found some interesting snippets on Pinal Dave’s website. Naturally, I had to take it further than the code snippet, HOWEVER it was a great place to start. Creating a temp table to hold database file details First off, I created a temp table which would hold the details of the individual data files within the database. Although there are a plethora of fields (within the temp table below), I utilize LogicalName only within this example. The temporary table structure may be seen below:   create table #tmp ( LogicalName nvarchar(128)  ,PhysicalName nvarchar(260)  ,Type char(1)  ,FileGroupName nvarchar(128)  ,Size numeric(20,0)  ,MaxSize numeric(20,0), Fileid tinyint, CreateLSN numeric(25,0), DropLSN numeric(25, 0), UniqueID uniqueidentifier, ReadOnlyLSN numeric(25,0), ReadWriteLSN numeric(25,0), BackupSizeInBytes bigint, SourceBlocSize int, FileGroupId int, LogGroupGUID uniqueidentifier, DifferentialBaseLSN numeric(25,0), DifferentialBaseGUID uniqueidentifier, IsReadOnly bit, IsPresent bit,  TDEThumbPrint varchar(50) )    We now declare and populate a variable(@path), setting the variable to the path to our SOURCE database backup. declare @path varchar(50) set @path = 'P:\DATA\MYDATABASE.bak'   From this point, we insert the file details of our database into the temp table. Note that we do so by utilizing a restore statement HOWEVER doing so in ‘filelistonly’ mode.   insert #tmp EXEC ('restore filelistonly from disk = ''' + @path + '''')   At this point, I depart from what I gleaned from Pinal Dave.   I now instantiate a few more local variables. The use of each variable will be evident within the cursor (which follows):   Declare @RestoreString as Varchar(max) Declare @NRestoreString as NVarchar(max) Declare @LogicalName  as varchar(75) Declare @counter as int Declare @rows as int set @counter = 1 select @rows = COUNT(*) from #tmp  -- Count the number of records in the temp                                    -- table   Declaring and populating the cursor At this point I do realize that many people are cringing about the use of a cursor. Being an Oracle professional as well, I have learnt that there is a time and place for cursors. I would remind the reader that the data that will be read into the cursor is from a local temp table and as such, any locking of the records (within the temp table) is not really an issue.   DECLARE MY_CURSOR Cursor  FOR  Select LogicalName  From #tmp   Parsing the logical names from within the cursor. A small caveat that works in our favour,  is that the first logical name (of our database) is the logical name of the primary data file (.mdf). Other files, except for the very last logical name, belong to secondary data files. The last logical name is that of our database log file.   I now open my cursor and populate the variable @RestoreString Open My_Cursor  set @RestoreString =  'RESTORE DATABASE [MYDATABASE] FROM DISK = N''P:\DATA\ MYDATABASE.bak''' + ' with  '   We now fetch the first record from the temp table.   Fetch NEXT FROM MY_Cursor INTO @LogicalName   While there are STILL records left within the cursor, we dynamically build our restore string. Note that we are using concatenation to create ‘one big restore executable string’.   Note also that the target physical file name is hardwired, as is the target directory.   While (@@FETCH_STATUS <> -1) BEGIN IF (@@FETCH_STATUS <> -2) -- As long as there are no rows missing select @RestoreString = case  when @counter = 1 then -- This is the mdf file    @RestoreString + 'move  N''' + @LogicalName + '''' + ' TO N’’X:\DATA1\'+ @LogicalName + '.mdf' + '''' + ', '   -- OK, if it passes through here we are dealing with an .ndf file -- Note that Counter must be greater than 1 and less than the number of rows.   when @counter > 1 and @counter < @rows then -- These are the ndf file(s)    @RestoreString + 'move  N''' + @LogicalName + '''' + ' TO N’’X:\DATA1\'+ @LogicalName + '.ndf' + '''' + ', '   -- OK, if it passes through here we are dealing with the log file When @LogicalName like '%log%' then    @RestoreString + 'move  N''' + @LogicalName + '''' + ' TO N’’X:\DATA1\'+ @LogicalName + '.ldf' +'''' end --Increment the counter   set @counter = @counter + 1 FETCH NEXT FROM MY_CURSOR INTO @LogicalName END   At this point we have populated the varchar(max) variable @RestoreString with a concatenation of all the necessary file names. What we now need to do is to run the sp_executesql stored procedure, to effect the restore.   First, we must place our ‘concatenated string’ into an nvarchar based variable. Obviously this will only work as long as the length of @RestoreString is less than varchar(max) / 2.   set @NRestoreString = @RestoreString EXEC sp_executesql @NRestoreString   Upon completion of this step, the database should be restored to the server. I now close and deallocate the cursor, and to be clean, I would also drop my temp table.   CLOSE MY_CURSOR DEALLOCATE MY_CURSOR GO   Conclusion Restoration of databases on different servers with different physical names and on different drives are a fact of life. Through the use of a few variables and a simple cursor, we may achieve an efficient and effective way to achieve this task.

    Read the article

  • Resolving data redundancy up front

    - by okeofs
    Introduction As all of us do when confronted with a problem, the resource of choice is to ‘Google it’. This is where the plot thickens. Recently I was asked to stage data from numerous databases which were to be loaded into a data warehouse. To make a long story short, I was looking for a manner in which to obtain the table names from each database, to ascertain potential overlap.   As the source data comes from a SQL database created from dumps of a third party product,  one could say that there were +/- 95 tables for each database.   Yes I know that first instinct is to use the system stored procedure “exec sp_msforeachdb 'select "?" AS db, * from [?].sys.tables'”. However, if one stops to think about this, it would be nice to have all the results in a temporary or disc based  table; which in itself , implies additional labour. This said,  I decided to ‘re-invent’ the wheel. The full code sample may be found at the bottom of this article.   Define a few temporary tables and variables   declare @SQL varchar(max); declare @databasename varchar(75) /* drop table ##rawdata3 drop table #rawdata1 drop table #rawdata11 */ -- A temp table to hold the names of my databases CREATE TABLE #rawdata1 (    database_name varchar(50) ,    database_size varchar(50),    remarks Varchar(50) )     --A temp table with the same database names as above, HOWEVER using an --Identity number (recNO) as a loop variable. --You will note below that I loop through until I reach 25 (see below) as at --that point the system databases, the reporting server database etc begin. --1- 24 are user databases. These are really what I was looking for. --Whilst NOT the best solution,it works and the code was meant as a quick --and dirty. CREATE TABLE #rawdata11 (    recNo int identity(1,1),    database_name varchar(50) ,    database_size varchar(50),    remarks Varchar(50) )   --My output table showing the database name and table name CREATE TABLE ##rawdata3 (    database_name varchar(75) ,    table_name varchar(75), )   Insert the database names into a temporary table I pull the database names using the system stored procedure sp_databases   INSERT INTO #rawdata1 EXEC sp_databases Go   Insert the results from #rawdata1 into a table containing a record number  #rawdata11 so that I can LOOP through the extract   INSERT into #rawdata11 select * from  #rawdata1   We now declare 3 more variables:  @kounter is used to keep track of our position within the loop. @databasename is used to keep track of the’ current ‘ database name being used in the current pass of the loop;  as inorder to obtain the tables for that database we  need to issue a ‘USE’ statement, an insert command and other related code parts. This is the challenging part. @sql is a varchar(max) variable used to contain the ‘USE’ statement PLUS the’ insert ‘ code statements. We now initalize @kounter to 1 .   declare @kounter int; declare @databasename varchar(75); declare @sql varchar(max); set @kounter = 1   The Loop The astute reader will remember that the temporary table #rawdata11 contains our  database names  and each ‘database row’ has a record number (recNo). I am only interested in record numbers under 25. I now set the value of the temporary variable @DatabaseName (see below) .Note that I used the row number as a part of the predicate. Now, knowing the database name, I can create dynamic T-SQL to be executed using the sp_sqlexec stored procedure (see the code in red below). Finally, after all the tables for that given database have been placed in temporary table ##rawdata3, I increment the counter and continue on. Note that I used a global temporary table to ensure that the result set persists after the termination of the run. At some stage, I plan to redo this part of the code, as global temporary tables are not really an ideal solution.    WHILE (@kounter < 25)  BEGIN  select @DatabaseName = database_name from #rawdata11 where recNo = @kounter  set @SQL = 'Use ' + @DatabaseName + ' Insert into ##rawdata3 ' + + ' SELECT table_catalog,Table_name FROM information_schema.tables' exec sp_sqlexec  @Sql  SET @kounter  = @kounter + 1  END   The full code extract   Here is the full code sample.   declare @SQL varchar(max); declare @databasename varchar(75) /* drop table ##rawdata3 drop table #rawdata1 drop table #rawdata11 */ CREATE TABLE #rawdata1 (    database_name varchar(50) ,    database_size varchar(50),    remarks Varchar(50) ) CREATE TABLE #rawdata11 (    recNo int identity(1,1),    database_name varchar(50) ,    database_size varchar(50),    remarks Varchar(50) ) CREATE TABLE ##rawdata3 (    database_name varchar(75) ,    table_name varchar(75), )   INSERT INTO #rawdata1 EXEC sp_databases go INSERT into #rawdata11 select * from  #rawdata1 declare @kounter int; declare @databasename varchar(75); declare @sql varchar(max); set @kounter = 1 WHILE (@kounter < 25)  BEGIN  select @databasename = database_name from #rawdata11 where recNo = @kounter  set @SQL = 'Use ' + @DatabaseName + ' Insert into ##rawdata3 ' + + ' SELECT table_catalog,Table_name FROM information_schema.tables' exec sp_sqlexec  @Sql  SET @kounter  = @kounter + 1  END    select * from ##rawdata3  where table_name like '%SalesOrderHeader%'

    Read the article

< Previous Page | 355 356 357 358 359 360 361 362 363 364 365 366  | Next Page >