Search Results

Search found 45542 results on 1822 pages for 'enable add ons'.

Page 383/1822 | < Previous Page | 379 380 381 382 383 384 385 386 387 388 389 390  | Next Page >

  • New Interoperability Solutions for SQL Server 2012

    - by The Official Microsoft IIS Site
    I am excited to share some great news about how we are opening up the SQL Server data platform even further with expanded interoperability support through new tools that allow customers to modernize their infrastructure while maximizing existing investments and extending virtually any data anywhere. The SQL Server team today introduced several tools that enable interoperability with SQL Server 2012. These tools help developers to build secure, highly available and high performance applications for...(read more)

    Read the article

  • Overview: Unique offerings for OPN members

    - by michaela.seika(at)oracle.com
    You need knowledge and skills to pass the exams to get specialised. We have mapped the Bootcamps and Courses that will enable you to do this. Oracle University knows that you need knowledge and skills quickly and recognises that you learn fast. Accelerate your learning curve by taking one of our OPN Only Bootcamps . They have highly attractive prices and your OPN discount is applied on top of this. View the schedule for each country at the following webpage:http://education.oracle.com/pls/web_prod-plq-dad/db_pages.getpage?page_id=400&p_name=OPNOFFERS

    Read the article

  • Interested in going to the cloud then this might be useful

    - by simonsabin
    Bob Duffy is doing an afternoon seminar on Azure. It will provide an introduction to the Azure platform, and in particular SQL Azure, show tools and methodologies to migrate on premise databases into the cloud, using a sample application and database and finally it will detail some of the Azure specific features that enable massive scale OLTP solutions such as federations. http://www.prodata.ie/Events/2012/SQL-Azure_and_the_Cloud.aspx...(read more)

    Read the article

  • &quot;CLR Enabled&quot; is not required to use CLR built-ins

    - by AaronBertrand
    Books Online articles referencing built-in CLR functions (such as FORMAT() ) have a remark similar to the following: "FORMAT relies on the presence of .the .NET Framework Common Language Runtime (CLR)." A lot of people seem to interpret this as meaning: "You must enable the sp_configure option 'CLR enabled' in order to use FORMAT()." Some then go on and suggest you run code similar to the following before you play with these functions: EXEC sp_configure 'show advanced options' , 1 ; GO RECONFIGURE...(read more)

    Read the article

  • How to install nuget package manually :)

    - by Anirudha
    Originally posted on: http://geekswithblogs.net/anirugu/archive/2013/11/13/how-to-install-nuget-package-manually.aspxWhen you install a nuget package in your project then Visual studio cached it for you for future purpose. If you live offline few time you can add the this cache directory as your nuget source. For see the folder where nuget cached then just go to Tools > options > Package manager > source > Cached packages > browse. You can add this cached folder as source.     For use nuget package without Visual studio you can try this trick.   open this in your explorer C:\Users\{yourusername}\AppData\Local\NuGet\Cache Now rename the file as described here. for example bootstrap.3.0.2.nupkg to bootstrap.3.0.2.rar open the rar and see the content folder. this folder contain the assembly/library that nuget install in your project.

    Read the article

  • Sempron 145 core unlocking

    - by Grant
    Call me a cheapskate looking for performance, but I have a Sempron 145 processor that has been unlocked to an Athlon II X2 4450e with a GIGABYTE GA-990FXA-UD3 motherboard, but only one of the two processor cores can be used at a time. If both of the cores are enabled in the BIOS the kernel will hang during boot, If I reboot the system and disable the second core (but not re-lock the processor), it will take me to the grub and then start normally. Is there a way to enable both cores and boot without kernel hangs?

    Read the article

  • Ubuntu 13.10 - Wifi not re-connecting after suspend

    - by Matthew
    I upgraded the other day from Ubuntu 13.04 to 13.10 on my ZaReason laptop, and the WiFi doesn't reconnect after suspend. In 13.04, pressing Fn+F2 turned network connections back on after suspend. The 'soft' wireless hardware switch. But in 13.10, that doesn't work, and trying to enable networking in the networking menu doesn't work either. So I have to restart every time. Any suggestions to make it work?

    Read the article

  • Security for LDAP authentication for Collabnet

    - by Robert May
    In a previous post, I wrote about how to get LDAP authentication working in Collabnet. By default, all LDAP users are put into the Users role on the server.  For most purposes, this is just fine, and I don’t have a way to change this.  The documentation gives hints that you can add them to other roles, but for now, I don’t have the need. However, adding permissions to different repositories is a different question. To add them, go to the repositories list, select Access Rules and then you can enter in their username, as it sits in Active Directory to the lists for the repositories or for the predefined groups that you have created.  To my knowledge, you cannot use the Active Directory groups in collabnet, which is a big problem.  Needing to micromanage users really limits the usefulness of the LDAP integration. Technorati Tags: subversion,collabnet

    Read the article

  • Speed up content loading

    - by user1806687
    I am using WinForms Sample downloaded from microsoft website. The problem is, that the model loading time is quite long, using: contentBuilder.Add(ModelPath, ModelName, null, "ModelProcessor"); contentManager.Load<Model>(ModelName); even a simple model, such as a cube with no textures, takes 4+ seconds to load. Now, I am no expert on this, but is there anyway to decrease loading time? EDIT: I've gone thru the code and found out that calling contentBuilder.Build(); ,which comes right after contentBuilder.Add() method takes up most of the time.

    Read the article

  • Fusion-Icon does not show up in system tray

    - by Schubibuntu
    I can't find a solution for my problem online: I'm running Unity 2D in Ubuntu 11.10 and want to enable compiz effects. I installed compiz config settings manager, and Fusion-Icon, but when I start Fusion-Icon nothing happens. When I execute it from terminal it stucks at lines like "setting update 'command'" or "setting update 'next_key'" I already set the whitelist for the panel to 'all'. Anyone knows what I'm missing? Thanks in advance!

    Read the article

  • Installing Ubuntu via Wubi - about SSDs and HDDs

    - by A Person
    I read the question in regards to installing Ubuntu with Wubi on an SSD ( How do I enable Trim on an SSD in a Wubi install under Windows 7? ). My Windows installation is on an SSD (C: drive). My question is, if I choose to let Wubi install Ubuntu on my D: drive (HDD), thus creating the root.disk-file on D:, would I need to worry about TRIM in any way? I would assume not, but I just want to be sure.

    Read the article

  • setup test domain ready for site launch

    - by nextyear
    I'm about to launch a site on a live server, after developing it with xampp on localhost. I first want to test the site before i make it live. How do i setup this up so I have it in a subdomain (i.e. test.livesite.com)? Is it just set it up on the server and only add the cname dns setting? Or is there a better way? All i am trying to do is add my site to the server, so i can edit it and look at it before I set it as live.

    Read the article

  • No windows preview option(Similar tp Windows 7) in compiz?

    - by harsh
    Guys I saw a video on the net regarding compiz to enable windows preview option when i hover a mouse over it. I installed compiz through Ubuntu Software centre and then I searched and found that there is no option called "Windows Preview" in it ! There is a unity option but it doesn't have a windows preview option. Also I would like the freedom to move unity anywhere from top-bottom-left-right. Will this damage my system? Please reply at the earliest :-|

    Read the article

  • Access Control Lists for Roles

    - by Kyle Hatlestad
    Back in an earlier post, I wrote about how to enable entity security (access control lists, aka ACLs) for UCM 11g PS3.  Well, there was actually an additional security option that was included in that release but not fully supported yet (only for Fusion Applications).  It's the ability to define Roles as ACLs to entities (documents and folders).  But now in PS5, this security option is now fully supported.   [Read More]

    Read the article

  • TechEd 2014 Day 3

    - by John Paul Cook
    There is some confusion about durability of data stored in SQL Server in-memory tables, so some review of the concepts is appropriate. The in-memory option is enabled at the database level. Enabling it at the database level only gives you the option to specify the in-memory feature on a table by table basis. No existing tables or new tables will by default become in-memory tables when you enable the feature at the database level. If you choose to make a table an in-memory table, by default it is...(read more)

    Read the article

  • Partial Function Application

    - by AngelEyes
    This is long overdo... Just a short and simple example of Partial Function Application.   For some good explanations, which also include the difference between Currying and Partial Function Application, check out: http://msmvps.com/blogs/jon_skeet/archive/2012/01/30/currying-vs-partial-function-application.aspx and also this answer on stackoverflow: http://stackoverflow.com/a/8826698/532517   And here is the example, taken from real code:           internal void Addfile(string parentDirName, string fileName, long size)         {             AddElement(parentDirName, fileName, (name, parent) => new File(name, size, parent));         }           public void AddDir(string parentDirName, string dirName)         {             AddElement(parentDirName, dirName, (name, parent) => new Directory(dirName, parent));         }           private void AddElement(string parentDirName, string name,                                 Func<string, FileSystemElement, FileSystemElement> elementCreator)         {             ValidateFileNames(parentDirName, name);             var parent = (Directory)_index[parentDirName];             FileSystemElement element = elementCreator(name, parent);             parent._elements.Add(element);             _index.Add(name, element);         }

    Read the article

  • Identifying Trends Using Web Analytics

    With a wide variety of web analytics tools at their disposal, marketers have the information they need to identify trends in the marketplace. More importantly, these tools enable marketers to make actionable insights in order to change or optimize various marketing functions and methods.

    Read the article

  • Deep Cloning C++ class that inherits CCNode in Cocos2dx

    - by A Devanney
    I stuck with something in Cocos2dx ... I'm trying to deep clone one of my classes that inherits CCNode. Basically i have.... GameItem* pTemp = new GameItem(*_actualItem); // loops through all the blocks in gameitem and updates their position pTemp->moveDown(); // if in boundary or collision etc... if (_gameBoard->isValidMove(pTemp)) { _actualItem = pTemp; // display the position CCLog("pos (1) --- (X : %d,Y : %d)", _actualItem->getGridX(),_actualItem->getGridY()); } Then doesn't work, because the gameitem inherits CCNode and has the collection of another class that also inherits CCNode. its just creating a shallow copy and when you look at children of the gameitem node in the copy, just point to the original? class GameItem : public CCNode { // maps to the actual grid position of the shape CCPoint* _rawPosition; // tracks the current grid position int _gridX, _gridY; // tracks the change if the item has moved CCPoint _offset; public: //constructors GameItem& operator=(const GameItem& item); GameItem(Shape shape); ... } then in the implementation.... GameItem& GameItem::operator=(const GameItem& item) { _gridX = item.getGridX(); _gridY = item.getGridY(); _offset = item.getOffSet(); _rawPosition = item.getRawPosition(); // how do i copy the node? return *this; } // shape contains an array of position for the game character GameItem::GameItem(Shape shape) { _rawPosition = shape.getShapePositions(); //loop through all blocks in position for (int i = 0; i < 7; i++) { // get the position of the first block in the shape and add to the position of the first block int x = (int) (getRawPosition()[i].x + getGridX()); int y = (int) (getRawPosition()[i].y + getGridY()); //instantiate a block with the position and type Block* block = Block::blockWithFile(x,y,(i+1), shape); // add the block to the this node this->addChild(block); } } And for clarity here is the block class class Block : public CCNode{ private: // using composition over inheritance CCSprite* _sprite; // tracks the current grid position int _gridX, _gridY; // used to store actual image number int _blockNo; public: Block(void); Block(int gridX, int gridY, int blockNo); Block& operator=(const Block& block); // static constructor for the creation of a block static Block* blockWithFile(int gridX, int gridY,int blockNo, Shape shape); ... } The blocks implementation..... Block& Block::operator=(const Block& block) { _sprite = new CCSprite(*block._sprite); _gridX = block._gridX; _gridY = block._gridY; _blockNo = block._blockNo; //again how to clone CCNode? return *this; } Block* Block::blockWithFile(int gridX, int gridY,int blockNo, Shape shape) { Block* block = new Block(); if (block && block->initBlockWithFile(gridX, gridY,blockNo, shape)) { block->autorelease(); return block; } CC_SAFE_DELETE(block); return NULL; } bool Block::initBlockWithFile(int gridX, int gridY,int blockNo, Shape shape) { setGridX(gridX); setGridY(gridY); setBlockNo(blockNo); const char* characterImg = helperFunctions::Format(shape.getFileName(),blockNo); // add to the spritesheet CCTexture2D* gameArtTexture = CCTextureCache::sharedTextureCache()->addImage("Character.pvr.ccz"); CCSpriteBatchNode::createWithTexture(gameArtTexture); // block settings _sprite = CCSprite::createWithSpriteFrameName(characterImg); // set the position of the block and add it to the layer this->setPosition(CONVERTGRIDTOACTUALPOS_X_Y(gridX,gridY)); this->addChild(_sprite); return true; } Any ideas are welcome at this point!! thanks

    Read the article

  • Update information outdated, "Failed to fetch cdrom"

    - by user285603
    I have a warning triangle on the top of my screen. When I click on it, it says that my update information is outdated. When I type sudo apt-get update && sudo apt-get upgrade into a terminal, I get this message: W: Failed to fetch cdrom://Ubuntu 14.04 LTS _Trusty Tahr_ - Release i386 (20140417)/dists/trusty/main/binary-i386/Packages Please use apt-cdrom to make this CD-ROM recognized by APT. apt-get update cannot be used to add new CD-ROMs W: Failed to fetch cdrom://Ubuntu 14.04 LTS _Trusty Tahr_ - Release i386 (20140417)/dists/trusty/restricted/binary-i386/Packages Please use apt-cdrom to make this CD-ROM recognized by APT. apt-get update cannot be used to add new CD-ROMs E: Some index files failed to download. They have been ignored, or old ones used instead. Any ideas?

    Read the article

  • Constrained A* problem

    - by Ragekit
    I've got a little problem with an A* algorithm that I need to Constrained a little bit. Basically : I use an A* to find the shortest path between 2 randomly placed room in 3D space, and then build a corridor between them. The problem I found is that sometimes it makes chimney like corridors that are not ideal, so I constrict the A* so that if the last movement was up or down, you go sideways. Everything is fine, but in some corner cases, it fails to find a path (when there is obviously one). Like here between the blue and red dot : (i'm in unity btw, but i don't think it matters) Here is the code of the actual A* (a bit long, and some redundency) while(current != goal) { //add stair up / stair down foreach(Node<GridUnit> test in current.Neighbors) { if(!test.Data.empty && test != goal) continue; //bug at arrival; if(test == goal && penul !=null) { Vector3 currentDiff = current.Data.bounds.center - test.Data.bounds.center; if(!Mathf.Approximately(currentDiff.y,0)) { //wanna drop on the last if(!coplanar(test.Data.bounds.center,current.Data.bounds.center,current.Data.parentUnit.bounds.center,to.Data.bounds.center)) { continue; } else { if(Mathf.Approximately(to.Data.bounds.center.x, current.Data.parentUnit.bounds.center.x) && Mathf.Approximately(to.Data.bounds.center.z, current.Data.parentUnit.bounds.center.z)) { continue; } } } } if(current.Data.parentUnit != null) { Vector3 previousDiff = current.Data.parentUnit.bounds.center - current.Data.bounds.center; Vector3 currentDiff = current.Data.bounds.center - test.Data.bounds.center; if(!Mathf.Approximately(previousDiff.y,0)) { if(!Mathf.Approximately(currentDiff.y,0)) { //you wanna drop now : continue; } if(current.Data.parentUnit.parentUnit != null) { if(!coplanar(test.Data.bounds.center,current.Data.bounds.center,current.Data.parentUnit.bounds.center,current.Data.parentUnit.parentUnit.bounds.center)) { continue; }else { if(Mathf.Approximately(test.Data.bounds.center.x, current.Data.parentUnit.parentUnit.bounds.center.x) && Mathf.Approximately(test.Data.bounds.center.z, current.Data.parentUnit.parentUnit.bounds.center.z)) { continue; } } } } } g = current.Data.g + HEURISTIC(current.Data,test.Data); h = HEURISTIC(test.Data,goal.Data); f = g + h; if(open.Contains(test) || closed.Contains(test)) { if(test.Data.f > f) { //found a shorter path going passing through that point test.Data.f = f; test.Data.g = g; test.Data.h = h; test.Data.parentUnit = current.Data; } } else { //jamais rencontré test.Data.f = f; test.Data.h = h; test.Data.g = g; test.Data.parentUnit = current.Data; open.Add(test); } } closed.Add (current); if(open.Count == 0) { Debug.Log("nothingfound"); //nothing more to test no path found, stay to from; List<GridUnit> r = new List<GridUnit>(); r.Add(from.Data); return r; } //sort open from small to biggest travel cost open.Sort(delegate(Node<GridUnit> x, Node<GridUnit> y) { return (int)(x.Data.f-y.Data.f); }); //get the smallest travel cost node; Node<GridUnit> smallest = open[0]; current = smallest; open.RemoveAt(0); } //build the path going backward; List<GridUnit> ret = new List<GridUnit>(); if(penul != null) { ret.Insert(0,to.Data); } GridUnit cur = goal.Data; ret.Insert(0,cur); do{ cur = cur.parentUnit; ret.Insert(0,cur); } while(cur != from.Data); return ret; You see at the start of the foreach i constrict the A* like i said. If you have any insight it would be cool. Thanks

    Read the article

  • Is it dangerous for me to give some of my Model classes Control-like methods?

    - by Pureferret
    In my personal project I have tried to stick to MVC, but I've also been made aware that sticking to MVC too tightly can be a bad thing as it makes writing awkward and forces the flow of the program in odd ways (i.e. some simple functions can be performed by something that normally wouldn't, and avoid MVC related overheads). So I'm beginning to feel justified in this compromise: I have some 'manager programs' that 'own' data and have some way to manipulate it, as such I think they'd count as both part of the model, and part of the control, and to me this feels more natural than keepingthem separate. For instance: One of my Managers is the PlayerCharacterManager that has these methods: void buySkill(PlayerCharacter playerCharacter, Skill skill); void changeName(); void changeRole(); void restatCharacter(); void addCharacterToGame(); void createNewCharacter(); PlayerCharacter getPlayerCharacter(); List<PlayerCharacter> getPlayersCharacter(Player player); List<PlayerCharacter> getAllCharacters(); I hope the mothod names are transparent enough that they don't all need explaining. I've called it a manager because it will help manage all of the PlayerCharacter 'model' objects the code creates, and create and keep a map of these. I may also get it to store other information in the future. I plan to have another two similar classes for this sort of control, but I will orchestrate when and how this happens, and what to do with the returned data via a pure controller class. This splitting up control between informed managers and the controller, as opposed to operating just through a controller seems like it will simplify my code and make it flow more. My question is, is this a dangerous choice, in terms of making the code harder to follow/test/fix? Is this somethign established as good or bad or neutral? I oculdn't find anything similar except the idea of Actors but that's not quite why I'm trying to do. Edit: Perhaps an example is needed; I'm using the Controller to update the view and access the data, so when I click the 'Add new character to a player button' it'll call methods in the controller that then go and tell the PlayerCharacterManager class to create a new character instance, it'll call the PlayerManager class to add that new character to the player-character map, and then it'll add this information to the database, and tell the view to update any GUIs effected. That is the sort of 'control sequence' I'm hoping to create with these manager classes.

    Read the article

  • SharePoint 2010 BDC Model Deployment Issue: The default web application could not be determined.

    Yesterday I tried to deploy a Business Data Connectivity Model project created in Visual Studio 2010 to my SharePoint 2010 test server (all RTM versions), but during the deployment of the solution, SharePoint threw my following error: Add Solution:  Adding solution 'BCSDemo2.wsp'...  Deploying solution 'BCSDemo2.wsp'...Error occurred in deployment step 'Add Solution': The default web application could not be determined. Set the SiteUrl property in feature BCSDemo2_Feature1 to the URL of...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • git workflow for separating commits

    - by gman
    Best practices with git (or any VCS for that matter) is supposed to be to have each commit do the smallest change possible. But, that doesn't match how I work at all. For example I recently I needed to add some code that checked if the version of a plugin to my system matched the versions the system supports. If not print a warning that the plugin probably requires a newer version of the system. While writing that code I decided I wanted the warnings to be colorized. I already had code that colorized error message so I edited that code. That code was in the startup module of one entry to the system. The plugin checking code was in another path that didn't use that entry point so I moved the colorization code into a separate module so both entry points could use it. On top of that, in order to test my plugin checking code works I need to go edit UI/UX code to make sure it tells the user "You need to upgrade". When all is said and done I've edited 10 files, changed dependencies, the 2 entry points are now both dependant on the colorization code, etc etc. Being lazy I'd probably just git add . && git commit -a the whole thing. Spending 10-15 minutes trying to manipulate all those changes into 3 to 6 smaller commits seems frustrating which brings up the question Are there workflows that work for you or that make this process easier? I don't think I can some how magically always modify stuff in the perfect order since I don't know that order until after I start modifying and seeing what comes up. I know I can git add --interactive etc but it seems, at least for me, kind of hard to know what I'm grabbing exactly the correct changes so that each commit is actually going to work. Also, since the changes are sitting in the current directory it doesn't seem like it would be easy to run tests on each commit to make sure it's going to work short of stashing all the changes. And then, if it were to stash and then run the tests, if I missed a few lines or accidentally added a few too many lines I have no idea how I'd easily recover from that. (as in either grab the missing lines from the stash and then put the rest back or take the few extra lines I shouldn't have grabbed and shove them into the stash for the next commit. Thoughts? Suggestions? PS: I hope this is an appropriate question. The help says development methodologies and processes

    Read the article

  • Looking For Premiere Onsite SEO Workshops?

    You can now stop wasting time and money chasing algorithms as you can now enjoy personalized SEO workshops which have been rated premiere worldwide and where you will get all your SEO questions answered by some of the most experienced and knowledgeable experts within your own local community. These workshops offer basic, intermediate and advanced SEO training which is held over two to five days and in addition there is a six month mentoring program that is included. These workshops will enable you to create your own website pages which will become visible to thousands of visitors.

    Read the article

< Previous Page | 379 380 381 382 383 384 385 386 387 388 389 390  | Next Page >