Search Results

Search found 28930 results on 1158 pages for 'sql ce'.

Page 861/1158 | < Previous Page | 857 858 859 860 861 862 863 864 865 866 867 868  | Next Page >

  • L'authentification forte par Token compromise ? Victime d'une cyber-attaque, RSA se fait dérober des informations sur sa technologie

    L'authentification forte par Token compromise ? La société RSA victime d'une cyber-attaque se fait dérober des informations sur sa technologie La société RSA est bien connue dans le monde de la sécurité d'entreprise. Elle fournit les fameux « Token », ces petits récepteurs qui permettent une authentification forte. [IMG]http://ftp-developpez.com/gordon-fowler/Token.gif[/IMG] Le principe est simple. Une identification « classique » repose sur un identifiant et un mot de passe. Une authentification forte repose, en plus, sur un numéro changeant en permanence (au moins une fois par minute). Ce numéro est envoyé en continue sur ...

    Read the article

  • SSRS 2008 R2 KPIs with bullet graphs

    Key Performance Indicators are typically displayed in a scorecard with stop light indicators, which are either red, amber or green light icons. The limitation for these kind of indicators is that you can see the actual and target values in two different fields as well as see the status of the KPI in red, amber or green color. If the user wants to figure out the thresholds associated with the KPI, these values are generally not visible. Further, representing the threshold values in the scorecard itself defeats the purpose of the scorecard. The scorecard should display the KPI's status in the most summarized form and use a minimal amount of space on the dashboard. In this tip we would look at how to address this issue.

    Read the article

  • You couldn't write it - Expired SA account

    - by GrumpyOldDBA
    This is the stuff of DBA nightmares ! email trail: Q. Can you reset the SA account on server XXXXX, we think it has expired and now no-one can work. Connect to Server: Surely no-one would set up a Server with an sa account which expires? Thankfully not. Find sa password and change connection to use SA account. Connect without issue. Me. Have checked Server and account is fine. A. Thanks that's great, you've fixed it we can all work now....(read more)

    Read the article

  • Why is my query soooooo slow?

    - by geekrutherford
    A stored procedure used in our production environment recently became so slow it cause the calling web service to begin timing out. When running the stored procedure in Query Analyzer it took nearly 3 minutes to complete.   The stored procedure itself does little more than create a small bit of dynamic SQL which calls a view with a where clause at the end.   At first the thought was that the query used within the view needed to be optimized. The query is quite long and therefore easy to jump to this conclusion.   Fortunately, after bringing the issue to the attention of a coworker they asked "is there a where clause, and if so, is there an index on the column(s) in it?" I had no idea and quickly said as much. A quick check on the table/column utilized in the where clause indicated indeed there was no index.   Before adding the index, and after admitting I am no SQL wiz, I checked the internet for info on the difference between clustered and non-clustered indexes. I found the following site quite helpful OdeToCode. After adding the non-clustered index on the column, the query that used to take nearly 3 minutes now takes 10 seconds! Ah, if only I'd thought to do this ahead of time!

    Read the article

  • Creating a Corporate Data Hub

    - by BuckWoody
    The Windows Azure Marketplace has a rich assortment of data and software offerings for you to use – a type of Software as a Service (SaaS) for IT workers, not necessarily for end-users. Among those offerings is the “Data Hub” – a  codename for a project that ironically actually does what the codename says. In many of our organizations, we have multiple data quality issues. Finding data is one problem, but finding it just once is often a bigger problem. Lots of departments and even individuals have stored the same data more than once, and in some cases, made changes to one of the copies. It’s difficult to know which location or version of the data is authoritative. Then there’s the problem of accessing the data. It’s fairly straightforward to publish a database, share or other location internally to store the data. But then you have to figure out who owns it, how it is controlled, and pass out the various connection strings to those who want to use it. And then you need to figure out how to let folks access the internal data externally – bringing up all kinds of security issues. Finally, in many cases our user community wants us to combine data from the internally sources with external data, bringing up the security, strings, and exploration features up all over again. Enter the Data Hub. This is an online offering, where you assign an administrator and data stewards. You import the data into the service, and it’s available to you - and only you and your organization if you wish. The basic steps for this service are to set up the portal for your company, assign administrators and permissions, and then you assign data areas and import data into them. From there you make them discoverable, and then you have multiple options that you or your users can access that data. You’re then able, if you wish, to combine that data with other data in one location. So how does all that work? What about security? Is it really that easy? And can you really move the data definition off to the Subject Matter Experts (SME’s) that know the particular data stack better than the IT team does? Well, nothing good is easy – but using the Data Hub is actually pretty simple. I’ll give you a link in a moment where you can sign up and try this yourself. Once you sign up, you assign an administrator. From there you’ll create data areas, and then use a simple interface to bring the data in. All of this is done in a portal interface – nothing to install, configure, update or manage. After the data is entered in, and you’ve assigned meta-data to describe it, your users have multiple options to access it. They can simply use the portal – which actually has powerful visualizations you can use on any platform, even mobile phones or tablets.     Your users can also hit the data with Excel – which gives them ultimate flexibility for display, all while using an authoritative, single reference for the data. Since the service is online, they can do this wherever they are – given the proper authentication and permissions. You can also hit the service with simple API calls, like this one from C#: http://msdn.microsoft.com/en-us/library/hh921924  You can make HTTP calls instead of code, and the data can even be exposed as an OData Feed. As you can see, there are a lot of options. You can check out the offering here: http://www.microsoft.com/en-us/sqlazurelabs/labs/data-hub.aspx and you can read the documentation here: http://msdn.microsoft.com/en-us/library/hh921938

    Read the article

  • How to change password schema for Dovecot user authentication for an already existing mail server

    - by deb_lrnr
    Hello, I have an email server setup on Debian Lenny with Postfix, Dovecot, SASL and MySQL. Currently, the password scheme in my dovecot-sql.conf file is set to: CRYPT default_pass_scheme = CRYPT I would like to globally change the scheme to something stronger like SSHA, or MD5-CRYPT and re-hash all passwords with SSHA. What is the best way to do this? The Dovecot wiki mentions how passwords that don't follow the default scheme defined in dovecot-sql.conf can be prefixed with "{ssha}password", but I couldn't see anything regarding changing an already-existing scheme to a new one for all passwords that are already in the database. Thanks for your help!

    Read the article

  • Le projet Quaero s'impose dans l'analyse des contenus multimédia, et apporte de nouvelles technologi

    Le projet Quaero s'impose dans l'analyse des contenus multimédia, et apporte de nouvelles technologies innovantes Le projet franco-allemand Quaero s'impose dans l'analyse des contenus multimédias depuis 2008. Ce programme de recherche, décrit comme un «*accélérateurde l'innovation en Europe*», est en pleine ébullition depuis deux ans. Financé à hauteur de 40 millions d'euros annuel, il a généré*230 publications,*19 distinctions*dans des conférences ; et ses laboratoires ont produit 45 nouvelles technologies.* Des industriels partenaires financent et portent certains projets de la recherche vers l'industrie. Certaines technologies sont donc déjà en service, comme par exemple celle développée par Exalead permettant de ...

    Read the article

  • Feedback from SQLBits 8

    - by Peter Larsson
    This years SQLBits occurred in Brighton. Although I didn’t have the opportunity to attend the full conference, I did a presentation at Saturday. Getting to Brighton was easy. Drove to Copenhagen airport at 0415, flew 0605 and arrived at Gatwick 0735. Then I took the direct train to Brighton and showed up at 0830, just one hour before presenting. This was the easy part. Getting home was much worse. Presentation ended at 1030 and I had to rush to the train station to get back to London, change to tube for Heathrow. Made it at the gate just 15 seconds before closing. That included a half mile run in the airport… Anyway, yesterday I got the feedback for my presentation. It does look good, especially since English is not my first language. This is the first graph Seems to be just halfway between conference average and best session. I can live with that. Second graph shows more detail about attendees voting. It also look acceptable. A wider spread for the 9’s, but it is an inevitable effect from how attendees percept the session. I did get a lot of 8’s and the lower grades in an descending order. The two people voting 4 and 5 didn’t say why they voted this so I don’t know how to remedy this. Third graph is about each category of votes.   Again, I find this acceptable. The Session abstract and Speaker’s knowledge seems to follow attendees expectations compared to conference average. I seem to have met the attendees expectations (and some more) for the other four categories, also compared to conference average. Since this did encourage me, I believe I will present some more at future meetings. I do have a new presentation about something all developers are doing every day but they may not know it. I will also cover this new topic in the next Deep Dives II book. Stay tuned! //Peter

    Read the article

  • Sortie de Sencha Cmd V3 qui propose une vision plus globale des outils de développement

    Sortie de Sencha Cmd V3 qui propose une vision plus globale des outils de développement. Il y a quelques semaines, l'équipe Sencha avait décidé de réorganiser son développement de Sencha Cmd autour d'un concept très intéressant : un framework "aware", un compilateur JavaScript. Ce SDK propose une série d'outils performants qui automatisent une grande partie de la création de vos applications. Ces outils s'ajoutent aux frameworks Ext JS (pour la partie JavaScript) et Sencha Touch (pour la partie mobiles). [IMG]http://cdn.sencha.io/img/20120918-sencha-cmd/cmd.png[/IMG] Les gains issus des toutes premières optimisations...

    Read the article

  • Le parti pirate vogue vers les législatives, et propose une campagne "libre et open-source", voterez-vous pour le parti en juin ?

    Le parti pirate vogue vers les législatives À un moment où tout le monde à les yeux rivés vers les élections présidentielles, les pirates regardent déjà plus loin vers l'horizon et scrutent les élections législatives de juin prochain. Leur objectif ? Obtenir leur premier siège à l'Assemblée nationale. Mais en premier lieu, le but du parti est de faire connaître leurs idées et d'amener des débats souvent ignorés par les autres partis. Siéger à l'Assemblée nationale serait l'idéal pour se faire entendre de la sphère politique. Mais, et c'est tout aussi important, les idées doivent aussi être entendues par les citoyens. C'est aussi dans ce but que les pirates ont mis en ligne un

    Read the article

  • Developpez.com passe la barre des 8000 followers sur Twitter : merci !

    Developpez.com passe la barre des 8000 followers Sur Twitter [IMG]http://windows.developpez.com/images/logo-twitter.png[/IMG] Et tous les 3 mois 1000 followers nous rejoignent. Alors que dire de plus ? A part que ce succès est entièrement dû à nos lecteurs et à nos membres toujours plus nombreux et toujours plus fidèles. Des lecteurs et des membres (vous quoi !) à qui nous adressons un immense : [IMG]http://ftp-developpez.com/gordon-fowler/Merci.jpg[/IMG] Création de woodleywonderworks sous licence CC Et pour ceux qui so...

    Read the article

  • General High-Level Assessment

    - by tcarper
    Guys and Gals, I've been tasked with a doozy of an assignment. The objective is something akin to "laying of hands" on several database servers which work in concert to provide data to various Web, Client-Server and Tablet-Sync'd distributed Client-Server programs. More specifically, I've been asked to come up with a "Maintenance Plan" which includes recommendations for future work to improve these machines' performance/reliability/security/etc. Might there be some good articles on teh interwebs ya'll could point me towards which would give me some good basis to start? Articles describing "These are the top 4 overarching categories and this is how you should proceed when drilling down on each of them" sort-of-thing would be fabulous. The Databases are all SQL 2005, however the compatibility level is 80 and they were originally created with ERwin based on SQL 6.5. The OSs are all Windows Server 2003. Thanks all! Tim

    Read the article

  • Speaking at PASS 2012 Summit in Seattle #sqlpass

    - by Marco Russo (SQLBI)
    I will deliver two sessions at the next PASS Summit 2012: one is title Inside DAX Query Plans and the other is Near Real-Time Analytics with xVelocity (without DirectQuery).These will be two sessions that require a lot of preparation and even if I have already much to say, I still have a long work to do this summer in order to go deeper in several details that I want to investigate for completing these sessions.I already look forward to come back in Seattle!In the meantime, you have to study SSAS Tabular and if you want to get a real jumpstart why not attending one of the next SSAS Tabular Workshop Online? We are working on more dates for this fall, but there are a few dates already scheduled.And, last but not least, the early Rough Cuts edition of our upcoming SSAS Tabular book is finally available here (really near to the final print)!

    Read the article

  • Hazlii cu politisti

    - by interesante
    Mor 6 politisti si se ancheteaza cazul:-Pai...Trei dintre ei erau cu barca pe lac si doi si-au aprins cate-o tigara.Unul si-a adus aminte ca a uitat sa stinga chibritul si a sarit in apa sa-l stinga si sa-necat.Al doilea uitase se-si stinga chistocul si a sarit si el si sa-necat.-Si al treilea?-Nu pornea barca si s-a dat jos s-o-mpinga si sa-necat.-Bine, dar ceilalti trei ?-Ei au murit la reconstituire.....Distreaza-te copios si cu jocuri flash de pe un site cu jocuri online.Doua sotii de politisti stau de vorba. Una zice:- Draga, sotul meu are post langa o florarie. Niciodata nu mi-a adus vreo floare...- Si ce? Al meu are post langa conservator. O conserva n-am vazut pana acum...

    Read the article

  • Wophone : le nouvel OS mobile chinois qui veut concurrencer Android et iOS

    Wophone : le nouvel OS mobile chinois Pour concurrencer Android et l'iPhone China Unicom, l'un des trois plus gros opérateurs de téléphonie sans fil de Chine, a annoncé qu'il envisageait de lancer de nouveaux terminaux mobile équipé de son propre système d'exploitation maison baptisé « Woppone». La plate-forme, en cours de développement, est basée sur Linux. Elle pourra tourner sur les Smartphones et sur les tablettes. Le but d'un tel OS est de permettre à China Unicom, d'après sa propre communication, de démocratiser sa jeune génération de réseau sans fil 3G. Grâce à ce nouvel OS, China Unicom espère développer des appareils sans fil 3G plus rapide et réduire le temps néc...

    Read the article

  • Terrible Performance with SATA Drives on Dell PowerEdge, steps to troubleshoot?

    - by Tom
    I had asked this question earlier and the question went missing so here it is again. Bought a DELL Poweredge 2950 to use as in-house QA Server. Disk performance is beyond terrible, 1000-4000 ms response time on the drive with our SQL Server database .mdf. Sql Server disk queue upwards of 300 at times. I'm a software guy, can anyone help me with steps to determine the issue? I don't know what RAID controller it has, how can I determine that? I'm speculating it could be BIOS issue. Perhaps the server used to have another kind of drive in it and when I added SATA the ??? buffer size is wrong??? Perhaps I chose wrong options (chose defaults) when setting up the RAID 1 arrays? I thought RAID 1 was a performance array?

    Read the article

  • Parent-child hierarchies and unary operators in PowerPivot

    - by Marco Russo (SQLBI)
    Alberto wrote an excellent post describing how to implement the Unary Operator feature (which is present in Analysis Services) in PowerPivot (there was a previous post about parent-child hierarchies, too). I have to say that the solution is not so easy to implement as in Analysis Services, but it just works and, from a practical point of view, it is not so difficult to implement if you understand how it works and accept its limitations (only sum and subtractions are supported). I think that many...(read more)

    Read the article

  • Le navigateur mobile de Firefox est disponible en version pre-Alpha, pour des tests sur les téléphon

    Le navigateur mobile de Firefox est disponible en version pre-Alpha, pour des tests sur les téléphones Android Depuis quelques semaines, une build précoce de Fennec circulait sur certains sites Internet. La version mobile de Firefox était disponible au téléchargement pour les téléphones Android. Cependant, cette mouture n'était pas officielle puisqu'offerte par un particulier et optimisée pour le Droid. Aujourd'hui, Mozilla pallie à ce manque et rend disponible une build pre-Alpha de Fennec, compatible avec au moins avec Droid et le Nexus One. Cette version est uniquement destinée à la réalisation de tests. Elle ne permet pas encore les mises à jour automatiques, mais Mozilla la trouve assez aboutie pour être décorti...

    Read the article

  • Is there a canonical source supporting "all-surrogates"?

    - by user61852
    Background The "all-PK-must-be-surrogates" approach is not present in Codd's Relational Model or any SQL Standard (ANSI, ISO or other). Canonical books seems to elude this restrictions too. Oracle's own data dictionary scheme uses natural keys in some tables and surrogate keys in other tables. I mention this because these people must know a thing or two about RDBMS design. PPDM (Professional Petroleum Data Management Association) recommend the same canonical books do: Use surrogate keys as primary keys when: There are no natural or business keys Natural or business keys are bad ( change often ) The value of natural or business key is not known at the time of inserting record Multicolumn natural keys ( usually several FK ) exceed three columns, which makes joins too verbose. Also I have not found canonical source that says natural keys need to be immutable. All I find is that they need to be very estable, i.e need to be changed only in very rare ocassions, if ever. I mention PPDM because these people must know a thing or two about RDBMS design too. The origins of the "all-surrogates" approach seems to come from recommendations from some ORM frameworks. It's true that the approach allows for rapid database modeling by not having to do much business analysis, but at the expense of maintainability and readability of the SQL code. Much prevision is made for something that may or may not happen in the future ( the natural PK changed so we will have to use the RDBMS cascade update funtionality ) at the expense of day-to-day task like having to join more tables in every query and having to write code for importing data between databases, an otherwise very strightfoward procedure (due to the need to avoid PK colisions and having to create stage/equivalence tables beforehand ). Other argument is that indexes based on integers are faster, but that has to be supported with benchmarks. Obviously, long, varying varchars are not good for PK. But indexes based on short, fix-length varchar are almost as fast as integers. The questions - Is there any canonical source that supports the "all-PK-must-be-surrogates" approach ? - Has Codd's relational model been superceded by a newer relational model ?

    Read the article

  • Investigating on xVelocity (VertiPaq) column size

    - by Marco Russo (SQLBI)
      In January I published an article about how to optimize high cardinality columns in VertiPaq. In the meantime, VertiPaq has been rebranded to xVelocity: the official name is now “xVelocity in-memory analytics engine (VertiPaq)” but using xVelocity and VertiPaq when we talk about Analysis Services has the same meaning. In this post I’ll show how to investigate on columns size of an existing Tabular database so that you can find the most important columns to be optimized. A first approach can be looking in the DataDir of Analysis Services and look for the folder containing the database. Then, look for the biggest files in all subfolders and you will find the name of a file that contains the name of the most expensive column. However, this heuristic process is not very optimized. A better approach is using a DMV that provides the exact information. For example, by using the following query (open SSMS, open an MDX query on the database you are interested to and execute it) you will see all database objects sorted by used size in a descending way. SELECT * FROM $SYSTEM.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS ORDER BY used_size DESC You can look at the first rows in order to understand what are the most expensive columns in your tabular model. The interesting data provided are: TABLE_ID: it is the name of the object – it can be also a dictionary or an index COLUMN_ID: it is the column name the object belongs to – you can also see ID_TO_POS and POS_TO_ID in case they refer to internal indexes RECORDS_COUNT: it is the number of rows in the column USED_SIZE: it is the used memory for the object By looking at the ration between USED_SIZE and RECORDS_COUNT you can understand what you can do in order to optimize your tabular model. Your options are: Remove the column. Yes, if it contains data you will never use in a query, simply remove the column from the tabular model Change granularity. If you are tracking time and you included milliseconds but seconds would be enough, round the data source column to the nearest second. If you have a floating point number but two decimals are good enough (i.e. the temperature), round the number to the nearest decimal is relevant to you. Split the column. Create two or more columns that have to be combined together in order to produce the original value. This technique is described in VertiPaq optimization article. Sort the table by that column. When you read the data source, you might consider sorting data by this column, so that the compression will be more efficient. However, this technique works better on columns that don’t have too many distinct values and you will probably move the problem to another column. Sorting data starting from the lower density columns (those with a few number of distinct values) and going to higher density columns (those with high cardinality) is the technique that provides the best compression ratio. After the optimization you should be able to reduce the used size and improve the count/size ration you measured before. If you are interested in a longer discussion about internal storage in VertiPaq and you want understand why this approach can save you space (and time), you can attend my 24 Hours of PASS session “VertiPaq Under the Hood” on March 21 at 08:00 GMT.

    Read the article

  • PowerPivot FILTER condition optimizations

    - by Marco Russo (SQLBI)
    In the comments of a recent post from Alberto Ferrari there was an interesting note about different performance related to the order of conditions in a FILTER call. I investigated about that and Jeffrey Wang has been so nice to give me some info about actual implementation that I can share on a blog post. First of all, an important disclaimer: PowerPivot is intended to make life easier, not requiring the user to think how to write the order of elements in a formula just to get better performance....(read more)

    Read the article

  • Understanding WCF Hosting

     WCF is a flagship product from  Microsoft for developing distributed application using SOA. Prior to WCF   traditional ASMX Web services were hosted only on Internet Information Services (IIS). The hosting options for WCF services are significantly enhanced from ... [Read Full Article]

    Read the article

  • Filtrage du Net : le Conseil National du Numérique s'oppose au Gouvernement et aux ayants droit

    Filtrage du Net : le Conseil National du Numérique s'oppose Au Gouvernement et aux ayants droit Le Gouvernement français entend mettre en place tous les moyens possibles pour lutter contre le « piratage ». Y compris des moyens que certains qualifieront de très limites comme le filtrage des réseaux. Ou la possibilité de fermer un site sans passer par une autorité juridique dès lors qu'il représente « un risque sérieux et grave d'atteinte à l'ordre public ». Définition vague, absence du juge, ce projet de décret n'a pas manqué de faire réagir le tout jeune Conseil national du numérique (CNN) qui rappelle à l'AFP que « toute mesure de blocage ne peut intervenir qu'au terme d'un déb...

    Read the article

< Previous Page | 857 858 859 860 861 862 863 864 865 866 867 868  | Next Page >