Search Results

Search found 34699 results on 1388 pages for 'big show 2011'.

Page 42/1388 | < Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >

  • What is the best HTML specification to be used as of Q1 2011?

    - by Rob McKinnon
    While developing a web application, what is the best spec to use? HTML4.01 HTML5 XHTML trans XHTML1.1 I was taught to use XHTML1.0 strict in uni and to avoid applet/iframe/tables(except in forms). I noticed that some deprecated tags are available in HTML5. Is it safe to code in HTML5? If so should I use target='', and the aforementioned tags? I have noticed that there are many alternatives to choose from including canvas, object. I have no preference, although Iframe tags are being dispensed from sources like Facebook/Google/etc. What would be the best avenue to take for Spec as of now(Feb 2011)?

    Read the article

  • Does attending the upcoming Devdays 2011 have some value for a resume?

    - by systempuntoout
    This fall I'm 99% going to London to attend the awesome Devdays 2011; I have many reasons to go there and some of them are: Professional stuff Great people Awesome topics Unicorns Passion London :) Obviously all the cool technologies that will be discussed are light years far from my daily work but useful for my side projects and maybe for some future employment. Now, to get to the point; a coworker said to me that he won't come with me because Devday London is expensive, and something expensive should reward you with a certificate, a certificate that could have some value to the eyes on an employer. Is he right? Do you think that attenting to this kind of event have some value on a resume? Should it be highlighted? Does it have any value for a future employer?

    Read the article

  • What new games are available for Ubuntu in 2011?

    - by Oli
    I know there have been other questions like "What native games are available?" and they often have issues because they turn into a never-ending list of every game ever released for Linux. But I'd like to know what's coming out this year. Good answers can include: A game that's coming out in 2011 A Linux port being released for games that might be older (eg Trine) As much information and as many screenshots and links as possible Few old games unless they're doing a major update that changes the game very significantly. One game per answer, add as much information as possible and work with each other to build a catalogue of awesome things to look forward to this year.

    Read the article

  • Will the MacBook Pro Early 2011 work better with Ubuntu than the Air Late 2010?

    - by AllanCaeg
    I got this late 2010 11" MacBook Air. I'm having issues with NVIDIA graphics, especially with GNOME Shell. I'm thinking about selling this to switch to the new MacBook Pro, particularly the entry level 13" (see specs here), because of the Intel HD Graphics 3000. I assume that it will be more FOSS-friendly. I just want to point out that there are non-negotiable reasons why I'm keeping an Apple hardware at the moment, so let's keep this on topic. Will the MacBook Pro Early 2011 work better with Ubuntu than the Air Late 2010? Any other factors than the graphics hardware? Should we expect better NVIDIA graphics anytime soon?

    Read the article

  • Autodesk Maya et Qt : la version 2011 est la première à se baser sur Qt, uniquement pour le meilleur

    Autodesk vient de lancer une nouvelle version d'un de ses produits phare, Maya 2011. Cette version est un peu spéciale : elle est sortie en 2010, certes, mais c'est aussi la toute première version de Maya à se baser sur Qt ! Dans tous les derniers films d'animation, jeux vidéo, et tout ce qui se fait en trois dimensions, il est plus que très probable que des parties ont été créées à l'aide d'Autodesk Maya, un des logiciels de modélisation 3D les plus répandus et les plus avancés depuis une dizaine d'années. Ce logiciel, compatible avec les trois plateformes principales, Windows, Linux et Mac OS X, est à même d'animer les rêves les plus fous de n'importe quel studio d'animation. Si, auparavant, quelques plug-ins étaient réalisés avec Qt afin d'être compatible...

    Read the article

  • Pwn2Own 2011 : BlackBerry et l'iPhone 4 vaincus lors du concours de hacking, les produits Google demeurent intouchés

    Pwn2Own 2011 : BlackBerry et l'iPhone 4 vaincus lors du concours de hacking, les produits Google demeurent intouchés Mise à jour du 11.03.2011 par Katleen Pour sa deuxième journée, le Pwn2Own a encore fait quelques victimes, mais en a aussi épargné certains. Les victimes potentielles de la journée étaient les systèmes d'exploitation mobile mais aussi Firefox (3.6), le navigateur qui n'avait pas été malmené hier. Seulement, Sam Dash, qui devait lui régler son compte, ne s'est pas présenté au concours. "Je ne peux pas écrire une code d'exploitation viable" pour ce challenge, s'est-il justifié. En revanche, l'iPhone 4 et le BlackBerry Torch sont tombés. Pour le s...

    Read the article

  • Emploi : Bilan mitigé pour l'année 2010 pour les postes IT, vivement 2011 ?

    Emploi : Bilan de l'année 2010 pour les postes IT, que prévoir en 2011 ? L'année 2010 touche à sa fin, aussi, avant l'avènement de l'année suivante, il est bon de faire un bilan des rebondissements qui ont animé la rubrique emploi de Developpez.com ces douze derniers mois. Premier constat : c'était mieux qu'en 2009, année catastrophique en raison des répercussions de la crise économique. Cette année, le marché s'est réveillé. Certes, la reprise est modeste, mais elle est là, et ça se voit. Le nombre d'embauches a augmenté par rapport à l'année précédente. Les recrutements ont connu une belle croissance : la note trimestrielle de conjoncture du marché de l'emploi des cadres diffusée fin octobre 2010 par...

    Read the article

  • TechDays TV : la Catch TV gratuite des TechDays 2011 pour revivre les moments forts du rendez-vous des professionnels informatiques

    TechDays TV : la Catch TV gratuite des TechDays 2011 Pour revivre les moments forts du rendez?vous des professionnels informatiques de Microsoft La cinquième édition des TechDays, l'évènement de Microsoft réunissant les développeurs, professionnels et décideurs IT, s'est achevée la semaine dernière. Cette édition a été marquée par une actualité riche et de nombreuses innovations. TechDays TV permettait ainsi aux personnes n'ayant pas eu l'occasion de se rendre sur place d'assister à plusieurs sessions et de suivre l'événement en direct. Aujourd'hui, Microsoft propose la rediffusion intégrale et gratuite des contenus de cette WebTV (conférences plénières, sélections des meil...

    Read the article

  • La RC de Internet Explorer 9 disponible pour la clôture des TechDays 2011, téléchargez-là et faîtes nous part de vos avis

    La RC de IE9 est disponible Avec Do Not Track, le support de WebM et met encore plus l'accent sur la sécurité Microsoft vient d'annoncer l'arrivée de la Release Candidate de IE9. Une nouvelle qui clôt en beauté les TechDays 2011, son salon dédiée cette année au Cloud Computing. Avec cette nouvelle étape, Internet Explorer 9 intègre toutes ses fonctionnalités finales. Il ne reste plus à l'équipe du projet qu'à corriger les bogues et les dernières imperfections avant la sortie officielle. Cette RC bénéficie d'un très léger toilettage de l'UI avec des onglets carrés (et non plus aux angles arrondies) qui peuvent à présent être placés en dessous d...

    Read the article

  • Imagine Cup 2011 : présentation de la catégorie Développement Embarqué du concours de Microsoft pour les étudiants, inscriptions jusqu'au 27 janvier

    Imagine Cup 2011 : présentation de la catégorie Développement Embarqué Du concours de Microsoft pour les étudiants, inscriptions jusqu'au 27 janvier La désormais célèbre compétition de développement organisée par Microsoft pour les étudiants touche aussi au développement embarqué. Une catégorie exigeante mais dont les résultats peuvent être impressionnants. Dans cette catégorie, le but est de développer une application sur une plateforme Windows CE qui aura elle-même été adaptée aux besoins des candidats organisés en équipe de 1 à 4 membres (la présence d'un mentor est conseillée). Rappelons que les version Embedded de Windows sont particulièrement conçues pour être modulables et ...

    Read the article

  • Big Visible Charts

    - by Robert May
    An important part of Agile is the concept of transparency and visibility. In proper functioning teams, stakeholders can look at any team at any time in the iteration or release and see how that team is doing by simply looking at what we call Big Visible Charts. If you’ve done Scrum, you’ve seen these charts. However, interpreting these charts can often be an art form. There are several different charts that can be useful. In this newsletter, I’ll focus on the Iteration Burndown and Cumulative Flow charts. I’ve included a copy of the spreadsheet that I used to create the charts, and if you don’t have a tool that creates them for you, you can use this spreadsheet to do so. Our preferred tool for managing Scrum projects is Rally. Rally creates all of these charts for you, saving you quite a bit of time. The Iteration Burndown and Cumulative Flow Charts This is the main chart that teams use. Although less useful to stakeholders, this chart is critical to the team and provides quite a bit of information to the team about how their iteration is going. Most charts are a combination of the charts below, so you may need to combine aspects of each section to understand what is happening in your iterations. Ideal Ah, isn’t that a pretty picture? Unfortunately, it’s also very unrealistic. I’ve seen iterations that come close to ideal, but never that match perfectly. If your iteration matches perfectly, chances are, someone is playing with the numbers. Reality is just too difficult to have a burndown chart that matches this exactly. Late Planning Iteration started, but the team didn’t. You can tell this by the fact that the real number of estimated hours didn’t appear until day two. In the cumulative flow, you can also see that nothing was defined in Day one and two. You want to avoid situations like this. You’ll note that the team had to burn faster than is ideal to meet the iteration because of the late planning. This often results in long weeks and days. Testing Starved Determining whether or not testing is starved is difficult without the cumulative flow. The pattern in the burndown could be nothing more that developers not completing stories early enough or could be caused by stories being too big. With the cumulative flow, however, you see that only small bites are in progress and stories were completed early, but testing didn’t start testing until the end of the iteration, and didn’t complete testing all stories in the iteration. When this happens, question whether or not your testing resources are sufficient for your team and whether or not acceptance is adequately defined. No Testing With this one, both graphs show the same thing; the team needs testers and testing! Without testing, what was completed cannot be verified to make sure that it is acceptable to the business. If you find yourself in this situation, review your testing practices and acceptance testing process and make changes today. Late Development With this situation, both graphs tell a story. In the top graph, you can see that the hours failed to burn down as quickly as the team expected. This could be caused by the team not correctly estimating their hours or the team could have had illness or some other issue that affected them. Often, when teams are tackling something that is more unknown, they’ll run into technical barriers that cause the burn down to happen slower than expected. In the cumulative flow graph, you can see that not much was completed in the first few days. This could be because of illness or technical barriers or simply poor estimation. Testing was able to keep up with everything that was completed, however. No Tool Updating When you see graphs that look like this, you can be assured that it’s because the team is not updating the tool that generates the graphs. Review your policy for when they are to update. On the teams that I run, I require that each team member updates the tool at least once daily. You should also check to see how well the team is breaking down stories into tasks. If they’re creating few large tasks, graphs can look similar to this. As a general rule, I never allow tasks, other than Unit Testing and Uncertainty, to be greater than eight hours in duration. Scope Increase I always encourage team members to enter in however much time they think they have left on a task, even if that means increasing the total amount of time left to do. You get a much better and more realistic picture this way. Increasing time remaining could explain the burndown graph, but by looking at the cumulative flow graph, we can see that stories were added to the iteration and scope was increased. Since planning should consume all of the hours in the iteration, this is almost always a bad thing. If the scope change happened late in the iteration and the hours remaining were well below the ideal burn, then increasing scope is probably o.k., but estimation needs to get better. However, with the charts above, that’s clearly not what happened and the team was required to do extra work to make the iteration. If you find this happening, your product owner and ScrumMasters need training. The team also needs to learn to say no. Scope Decrease Scope decreases are just as bad as scope increases. Usually, graphs above show that the team did a poor job of estimating their stories and part way through had to reduce scope to change the iteration. This will happen once in a while, but if you find it’s a pattern on your team, you need to re-evaluate planning. Some teams are hopelessly optimistic. In those cases, I’ll introduce a task I call “Uncertainty.” With Uncertainty, the team estimates how many hours they might need if things don’t go well with the tasks they’ve defined. They try to estimate things that could go poorly and increase the time appropriately. Having an Uncertainty task allows them to have a low and high estimate. Uncertainty should not just be an arbitrary buffer. It must correlate to real uncertainty in the tasks that have been defined. Stories are too Big Often, we see graphs like the ones above. Note that the burndown looks fairly good, other than the chunky acceptance of stories. However, when you look at cumulative flow, you can see that at one point, everything is in progress. This is a bad thing. When you see graphs like this, you’re in one of two states. You may just have a very small team and can only handle one or two stories in your iteration. If you have more than one or two people, then the most likely problem is that your stories are far too big. To combat this, break large high hour stories into smaller pieces that can be completed independently and accepted independently. If you don’t, you’ll likely be requiring your testers to do heroic things to complete testing on the last day of the iteration and you’re much more likely to have the entire iteration fail, because of the limited amount of things that can be completed. Summary There are other charts that can be useful when doing scrum. If you don’t have any big visible charts, you really need to evaluate your process and change. These charts can provide the team a wealth of information and help you write better software. If you have any questions about charts that you’re seeing on your team, contact me with a screen capture of the charts and I’ll tell you what I’m seeing in those charts. I always want this information to be useful, so please let me know if you have other questions. Technorati Tags: Agile

    Read the article

  • Upcoming Upgrade Workshops in the US

    - by Mike Dietrich
    As Roy is really busy in traveling the whole North American continent I would like to highlight a few of Roy's upcoming workshops with registration links - so simply "click" and register :-) March 23, 2011: Philadelphia, PA March 24, 2011: Reston, VA April 07, 2011: Dallas, TX April 13, 2011: Birmingham, AL April 14, 2011: Minneapolis, MN Roy is looking forward to meet you in one of the above or the upcoming events in California and Oregon. Mike

    Read the article

  • Multiple vulnerabilities in OpenSSL

    - by chandan
    CVE DescriptionCVSSv2 Base ScoreComponentProduct and Resolution CVE-2011-4108 Cryptographic Issues vulnerability 4.3 OpenSSL Solaris 11 11/11 SRU 4a CVE-2011-4109 Unspecified vulnerability 9.3 CVE-2011-4576 Information Disclosure vulnerability 5.0 CVE-2011-4577 Denial of Service (DoS) vulnerability 4.3 CVE-2011-4619 Denial of Service (DoS) vulnerability 5.0 CVE-2012-0027 Denial of Service (DoS) vulnerability 5.0 This notification describes vulnerabilities fixed in third-party components that are included in Sun's product distribution.Information about vulnerabilities affecting Oracle Sun products can be found on Oracle Critical Patch Updates and Security Alerts page.

    Read the article

  • Outer Join is not working in Linq Query: The method 'Join' cannot follow the method 'SelectMany' or is not supported

    - by Scorpion
    I am writing the Linq query as below: But on run its throwing the following error: The method 'Join' cannot follow the method 'SelectMany' or is not supported. Try writing the query in terms of supported methods or call the 'AsEnumerable' or 'ToList' method before calling unsupported methods. LINQ from a in AccountSet join sm in new_schoolMemberSet on a.AccountId equals sm.new_OrganisationId.Id into ps from suboc in ps.DefaultIfEmpty() join sr in new_schoolRoleSet on suboc.new_SchoolRoleId.Id equals sr.new_schoolRoleId where sr.new_name == "Manager" where a.new_OrganisationType.Value == 430870007 select new { a.AccountId, a.new_OrganisationType.Value } I am expecting the result as below: I never used the Outer join in Linq before. So please correct me if I am doing it wrong. Thanks

    Read the article

  • How do I mix functions in complex SSRS expressions?

    - by Boydski
    I'm writing a report against a data repository that has null values within some of the columns. The problem is building expressions is as temperamental as a hormonal old lady and doesn't like my mixing of functions. Here's an expression I've written that does not work if the data in the field is null/nothing: =IIF( IsNumeric(Fields!ADataField.Value), RunningValue( IIF( DatePart("q", Fields!CreatedOn.Value) = "2", Fields!ADataField.Value, 0 ), Sum, Nothing ), Sum(0) ) (Pseudocode) "If the data is valid and if the data was created in the second quarter of the year, add it to the overall Sum, otherwise, add zero to the sum." Looks pretty straight forward. And the individual pieces of the expression work by themselves. IE: IsNumeric(), DatePart(), etc. But when I put them all together, the expression throws an error. I've attempted about every permutation of what's shown above, all to no avail. Null values in Fields!ADataField.Value cause errors. Thoughts?

    Read the article

  • How do you configure an SDL Tridion CME extension for a subset of views?

    - by Chris Summers
    I have created a new editor for SDL Tridion which adds some new functionality to the ribbon bar. This is enabled by adding the following snippet to the editor.config <!-- ItemCommenting PowerTool --> <ext:extension assignid="ItemCommenting" name="Save and&lt;br/&gt;Comment" pageid="HomePage" groupid="ManageGroup" insertbefore="SaveCloseBtn"> <ext:command>PT_ItemCommenting</ext:command> <ext:title>Save and Comment</ext:title> <ext:issmallbutton>false</ext:issmallbutton> <ext:dependencies> <cfg:dependency>PowerTools.Commands</cfg:dependency> </ext:dependencies> <ext:apply> <ext:view name="*" /> </ext:apply> </ext:extension> This is applied to all views by using a wildcard value in the node. This has results in my new button being added to the ribbon of every view, including the main dashboard. Is there a way to add this to all views except for the dashboard? Or do I have to create something like this? <ext:apply> <ext:view name="PageView" /> <ext:view name="ComponentView" /> <ext:view name="SchemaView" /> </ext:apply> If this is the only way to achieve the result I need, is there a list of all the view names somewhere?

    Read the article

  • Campaignresponse entity create in plugin

    - by engcmreng
    I wrote plugin that fire on phonecall step is setstatedynamic. I register post image in step my plugin.I get phonecall entity from postimage. And When i set phonecallid to origanitingactivityid on campaign response ( like campaignResponse.OriginatingActivityId = new EntityReference(); campaignResponse.OriginatingActivityId.LogicalName = PhoneCall.EntityLogicalName; campaignResponse.OriginatingActivityId.Id = phoneCall.ActivityId.Value; ) Plugin throws timeout.(I set serviceproxy,Timeout property 30 minutes,its still throw timeout.) When i dont set originatingactivityid,campaingresponse is created. Any suggestion about this problem.

    Read the article

  • Access container page of a component in .net based CT

    - by Huston Lopes
    We are developing a .net based CT based on Broker query Mechanism (filter): ComponentPresentationAssembler cpAssembler = new ComponentPresentationAssembler(Page ID,Page object); In order to pass the page ID, I need to get the access of page on which the component is present. How can I access the page from package? Since this a CT, a component object would be available in page and not a page object. Tried the following piece of code, but without success: string pageURI = _package.GetValue("Page.ID"); Page objPage = (Page)_engine.GetSession().GetObject(pageURI); This is not working as there is no page object. What are the alternatives so that we can access the parent page of component from CT?

    Read the article

  • How can I set a field to null in a Pre-Update Plugin

    - by Juergen
    I developed a Pre-Update Plugin for the Case entity. In this plugin I want to set a string field to a new value. It works smoothly if the new value is not null. But if the new value si null, it is just ignored. This works: Incident caseTarget = ((Entity) localContext.PluginExecutionContext.InputParameters["Target"]).ToEntity<Incident>(); caseTarget.ProductSerialNumber = "new value"; After the execution of the plugin, the ProductSerialNumber field has value "new value". This doesn't work: Incident caseTarget = ((Entity) localContext.PluginExecutionContext.InputParameters["Target"]).ToEntity<Incident>(); caseTarget.ProductSerialNumber = null; After the execution of the plugin, the ProductSerialNumber field has still its old value. How can I set the target's field to null?

    Read the article

  • Is it possible to publish component revisions within workflow?

    - by Ianthe
    Our current setup is that we have two targets, Staging and Live. Collaborators may update the affected component while it is still within the workflow. A final activity is set to publish the related pages to Live. Is it possible to publish component updates (revisions e.g. 2.2, 2.5) to Staging from within the workflow? The TOM API documentation for Page.Publish() method does not seem to have an input parameter to fullfil such purpose.

    Read the article

  • How do I download an attachment from an annotation using client-side JScript?

    - by VVander
    I'm trying to provide a link to the attachment of a note through the client-side JScript. The standard MS-made Notes component does this through the following url: [serverurl]/[appname]/Activities/Attachment/download.aspx?AttachmentType=5&AttachmentId={blahblahblah}&IsNotesTabAttachment=1&CRMWRPCToken=blahblahblah&CRMWRPCTokenTimeStamp=blahblahblah The problem is that I don't know how to get the Token or TokenTimeStamp, so I'm receiving an Access Denied error ("form is no longer available, security precaution, etc"). The only other way I can think of doing this is through the OData endpoint, but that would at best get me a base64 string that I still would have translate into a filestream to give to the browser (all of which seems like it would take forever to implement/figure out). I've found a few other posts that describe the same thing, but no one has answered them: http://social.microsoft.com/Forums/en-US/crmdevelopment/thread/6eb9e0d4-0c0c-4769-ab36-345fbfc9754f/ http://social.microsoft.com/Forums/is/crm/thread/45dabb6e-1c6c-4cb4-85a4-261fa58c04da

    Read the article

  • How to auto-generate early bound properties for Entity specific (ie Local) Option Set text values?

    - by Daryl
    After spending a year working with the Microsoft.Xrm.Sdk namespace, I just discovered yesterday the Entity.FormattedValues property contains the text value for Entity specific (ie Local) Option Set texts. The reason I didn't discover it before, is there is no early bound method of getting the value. i.e. entity.new_myOptionSet is of type OptionSetValue which only contains the int value. You have to call entity.FormattedValues["new_myoptionset"] to get the string text value of the OptionSetValue. Therefore, I'd like to get the crmsrvcutil to auto-generate a text property for local option sets. i.e. Along with Entity.new_myOptionSet being generated as it currently does, Entity.new_myOptionSetText would be generated as well. I've looked into the Microsoft.Crm.Services.Utility.ICodeGenerationService, but that looks like it is mostly for specifying what CodeGenerationType something should be... Is there a way supported way using CrmServiceUtil to add these properties, or am I better off writing a custom app that I can run that can generate these properties as a partial class to the auto-generated ones? Edit - Example of the code that I would like to be generated Currently, whenever I need to access the text value of a OptionSetValue, I use this code: var textValue = OptionSetCache.GetText(service, entity, e => e.New_MyOptionSet); The option set cache will use the entity.LogicalName, and the property expression to determine the name of the option set that I'm asking for. It will then query the SDK using the RetrieveAttriubteRequest, to get a list of the option set int and text values, which it then caches so it doesn't have to hit CRM again. It then looks up the int value of the New_MyOptionSet of the entity and cross references it with the cached list, to get the text value of the OptionSet. Instead of doing all of that, I can just do this (assuming that the entity has been retrieved from the server, and not just populated client side): var textValue = entity.FormattedValues["new_myoptionset"]; but the "new_myoptionset" is no longer early bound. I would like the early bound entity classes that gets generated to also generate an extra "Text" property for OptionSetValue properties that calls the above line, so my entity would have this added to it: public string New_MyOptionSetText { return this.GetFormattedAttributeValue("new_myoptionset"); // this is a protected method on the Entity class itself... }

    Read the article

  • CRM2011: Home page ribbon enable rule doesn't work properly

    - by nixjojo
    In my lead home page, there is a custom button. The enable rule for that button is: <EnableRule Id="enableruleid"> <SelectionCountRule AppliesTo="SelectedEntity" Minimum="1"></SelectionCountRule> <CustomRule FunctionName="functionname" Library="$Webresource:myjavascript.js"> <CrmParameter Value="SelectedControlSelectedItemIds" /> </CustomRule> </EnableRule> The javascript working fine only for the first time select a record, when you select another record, the javascript doesn't called. For example, I select record A, the button is enabled and it's fine; and then I select record B, the button should be disabled, but it's not, it still enabled. But if I select record B first, the ribbon is disabled as I wish, and then I select record A, the button still disabled.

    Read the article

  • Windows Azure Recipe: Big Data

    - by Clint Edmonson
    As the name implies, what we’re talking about here is the explosion of electronic data that comes from huge volumes of transactions, devices, and sensors being captured by businesses today. This data often comes in unstructured formats and/or too fast for us to effectively process in real time. Collectively, we call these the 4 big data V’s: Volume, Velocity, Variety, and Variability. These qualities make this type of data best managed by NoSQL systems like Hadoop, rather than by conventional Relational Database Management System (RDBMS). We know that there are patterns hidden inside this data that might provide competitive insight into market trends.  The key is knowing when and how to leverage these “No SQL” tools combined with traditional business such as SQL-based relational databases and warehouses and other business intelligence tools. Drivers Petabyte scale data collection and storage Business intelligence and insight Solution The sketch below shows one of many big data solutions using Hadoop’s unique highly scalable storage and parallel processing capabilities combined with Microsoft Office’s Business Intelligence Components to access the data in the cluster. Ingredients Hadoop – this big data industry heavyweight provides both large scale data storage infrastructure and a highly parallelized map-reduce processing engine to crunch through the data efficiently. Here are the key pieces of the environment: Pig - a platform for analyzing large data sets that consists of a high-level language for expressing data analysis programs, coupled with infrastructure for evaluating these programs. Mahout - a machine learning library with algorithms for clustering, classification and batch based collaborative filtering that are implemented on top of Apache Hadoop using the map/reduce paradigm. Hive - data warehouse software built on top of Apache Hadoop that facilitates querying and managing large datasets residing in distributed storage. Directly accessible to Microsoft Office and other consumers via add-ins and the Hive ODBC data driver. Pegasus - a Peta-scale graph mining system that runs in parallel, distributed manner on top of Hadoop and that provides algorithms for important graph mining tasks such as Degree, PageRank, Random Walk with Restart (RWR), Radius, and Connected Components. Sqoop - a tool designed for efficiently transferring bulk data between Apache Hadoop and structured data stores such as relational databases. Flume - a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large log data amounts to HDFS. Database – directly accessible to Hadoop via the Sqoop based Microsoft SQL Server Connector for Apache Hadoop, data can be efficiently transferred to traditional relational data stores for replication, reporting, or other needs. Reporting – provides easily consumable reporting when combined with a database being fed from the Hadoop environment. Training These links point to online Windows Azure training labs where you can learn more about the individual ingredients described above. Hadoop Learning Resources (20+ tutorials and labs) Huge collection of resources for learning about all aspects of Apache Hadoop-based development on Windows Azure and the Hadoop and Windows Azure Ecosystems SQL Azure (7 labs) Microsoft SQL Azure delivers on the Microsoft Data Platform vision of extending the SQL Server capabilities to the cloud as web-based services, enabling you to store structured, semi-structured, and unstructured data. See my Windows Azure Resource Guide for more guidance on how to get started, including links web portals, training kits, samples, and blogs related to Windows Azure.

    Read the article

< Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >