Search Results

Search found 30085 results on 1204 pages for 'read only'.

Page 305/1204 | < Previous Page | 301 302 303 304 305 306 307 308 309 310 311 312  | Next Page >

  • Is return-type-(only)-polymorphism in Haskell a good thing?

    - by dainichi
    One thing that I've never quite come to terms with in Haskell is how you can have polymorphic constants and functions whose return type cannot be determined by their input type, like class Foo a where foo::Int -> a Some of the reasons that I do not like this: Referential transparency: "In Haskell, given the same input, a function will always return the same output", but is that really true? read "3" return 3 when used in an Int context, but throws an error when used in a, say, (Int,Int) context. Yes, you can argue that read is also taking a type parameter, but the implicitness of the type parameter makes it lose some of its beauty in my opinion. Monomorphism restriction: One of the most annoying things about Haskell. Correct me if I'm wrong, but the whole reason for the MR is that computation that looks shared might not be because the type parameter is implicit. Type defaulting: Again one of the most annoying things about Haskell. Happens e.g. if you pass the result of functions polymorphic in their output to functions polymorphic in their input. Again, correct me if I'm wrong, but this would not be necessary without functions whose return type cannot be determined by their input type (and polymorphic constants). So my question is (running the risk of being stamped as a "discussion quesion"): Would it be possible to create a Haskell-like language where the type checker disallows these kinds of definitions? If so, what would be the benefits/disadvantages of that restriction? I can see some immediate problems: If, say, 2 only had the type Integer, 2/3 wouldn't type check anymore with the current definition of /. But in this case, I think type classes with functional dependencies could come to the rescue (yes, I know that this is an extension). Furthermore, I think it is a lot more intuitive to have functions that can take different input types, than to have functions that are restricted in their input types, but we just pass polymorphic values to them. The typing of values like [] and Nothing seems to me like a tougher nut to crack. I haven't thought of a good way to handle them. I doubt I am the first person to have had thoughts like these. Does anybody have links to good discussions about this Haskell design decision and the pros/cons of it?

    Read the article

  • Adding JavaScript to your code dependent upon conditions

    - by DavidMadden
    You might be in an environment where you code is source controlled and where you might have build options to different environments.  I recently encountered this where the same code, built on different configurations, would have the website at a different URL.  If you are working with ASP.NET as I am you will have to do something a bit crazy but worth while.  If someone has a more efficient solution please share. Here is what I came up with to make sure the client side script was placed into the HEAD element for the Google Analytics script.  GA wants to be the last in the HEAD element so if you are doing others in the Page_Load then you should do theirs last. The settings object below is an instance of a class that holds information I collection.  You could read from different sources depending on where you stored your unique ID for Google Analytics. *** This has been formatted to fit this screen. *** if (!IsPostBack) { if (settings.GoogleAnalyticsID != null || settings.GoogleAnalyticsID != string.Empty) { string str = @"//<!CDATA[ var _gaq = _gaq || []; _gaq.push(['_setAccount', '"  + settings.GoogleAnalyticsID + "']); _gaq.push(['_trackPageview']);  (function () {  var ga = document.createElement('script');  ga.type = 'text/javascript';  ga.async = true;  ga.src = ('https:' == document.location.protocol  ? 'https://ssl' :  'http://www') + '.google-analytics.com/ga.js'; var s = document.getElementsByTagName('script')[0];  s.parentNode.insertBefore(ga, s);})();"; System.Web.UI.HtmlControls.HtmlGenericControl si =  new System.Web.UI.HtmlControls.HtmlGenericControl(); si.TagName = "script"; si.Attributes.Add("type", @"text/javascript"); si.InnerHtml = sb.ToString(); this.Page.Header.Controls.Add(si); } } The code above will prevent the code from executing if it is a PostBack and then if the ID was not able to be read or something caused the settings to be lost by accident. If you have larger function to declare, you can use a StringBuilder to separate the lines. This is the most compact I wished to go and manage readability.

    Read the article

  • TPC-H Benchmarks on SQL Server 2014 with Columnstore

    - by jchang
    Three TPC-H benchmark results were published in April of this year at SQL Server 2014 launch, where the new updateable columnstore feature was used. SQL Server 2012 had non-updateable columnstore that required the base table to exist in rowstore form. This was not used in the one published TPC-H benchmark result on SQL Server 2012, which includes two refresh stored procedures, one inserting rows, the second deleting rows. It is possible that the TPC-H rules do not allow a view to union two tables?...(read more)

    Read the article

  • BonaVista Dimensions used as a report service

    - by Marco Russo (SQLBI)
    Recently I have seen a long demo of BonaVista Dimensions . It is a product that is able to create reports and, most important dashboards. You can use it also without SQL Server and Analysis Services, just by importing data in a local cube file that you can model using its own simple to use user interface. But what is interesting to me (in this post) is the capability to connect to a SSAS cube. It is somewhat similar to XLCubed and in reality these two products have something in common, because both...(read more)

    Read the article

  • How to Stop Users From Uninstalling Metro Applications in Windows 8

    - by Taylor Gibb
    We have already shown you how you can disable the Windows Store in Windows 8, but what if you have already installed some Metro applications or intend on doing so, but don’t want them to be uninstalled? Read on to find out how you can do this. How to Banish Duplicate Photos with VisiPic How to Make Your Laptop Choose a Wired Connection Instead of Wireless HTG Explains: What Is Two-Factor Authentication and Should I Be Using It?

    Read the article

  • SQL Server 2008 R2 Reporting Services: A generic error occurred in GDI+

    - by Greg Low
    While building some maps today in SQL Server 2008 R2 Reporting Services, I kept coming up with an error that said: A generic error occurred in GDI+ I was struggling to think what I'd done wrong. After much nashing of teeth and removal of hair, I finally worked out what the error was. When I got to the "Choose Color Scheme and Visualization" page of the map wizard for building a color analytical map, I hadn't picked the correct value for the "Field to Visualize" drop-down. I'm guessing that because...(read more)

    Read the article

  • I'm Speaking @ SQL in the City (London 15th July)

    - by NeilHambly
    If you didn't already know Redgate have 2 full 1 day conferences planned Called " SQL in the City ", these being held @ the following 2 major cities London, UK on 15th July (Now Full with waiting list) Los Angeles, US on 28th Oct It is a full days’ worth of FREE SQL Server training sponsored by Redgate, you get the opportunity to attend a number of training sessions in the track of your choosing, along with the chance to network with your peers and interact with SQL MVP's, Redgate...(read more)

    Read the article

  • Software Center doesn't show files installed

    - by postfuturist
    When I install a program in the Software Center that happens to install a menu option, it is discoverable through the menu, but if it is a package that is documentation, or command-line only, there is no simple way to discover how to actually access the content / software. Is there a way in the Software Center to see what files or executables were installed with a particular package so I can actually use / read it?

    Read the article

  • Facebook C# SDK submitted to the Outercurve Foundation

    - by The Official Microsoft IIS Site
    I am pleased to announce another open source milestone as we continue to deliver on our commitment to Interoperability: today, the Facebook C# SDK was submitted to the Outercurve Foundation’s Data, Languages, and Systems Interoperability gallery. This project is a set of libraries that enables developers of all Microsoft platforms, as well as Mono, to build applications that integrate with Facebook. The project contains core libraries for authentication and calling Facebook APIs. Additionally...(read more)

    Read the article

  • Great Solaris 10 features paving the way to Solaris 11

    - by Larry Wake
    Karoly Vegh writes on the Oracle Systems Blog Austria about what you can do with Solaris 10 today that will get you ready for Solaris 11. Even today, many people still use Solaris 10 as if it were a patch update to Solaris 8 or 9, missing out on the power behind Live Upgrade, Zones, resource management, and ZFS. Learning more about these will help set your feet on the road to the even more sophisticated capabilities of Oracle Solaris 11. [Read More]

    Read the article

  • Shared Datasets in SQL Server 2008 R2

    This article leverages the examples and concepts explained in the Part I through Part IV of the spatial data series which develops a "BI-Satellite" app. Overview In the spatial data series we ... [Read Full Article]

    Read the article

  • POST and PUT requests – is it just the convention?

    - by bckpwrld
    I've read quite a few articles on the difference between POST and PUT and in when the two should be used. But there are still few things confusing me ( hopefully questions will make some sense ): 1) We should use PUT to create resources when we want clients to specify the URI of the newly created resources and we should use POST to create resources when we let service generate the URI of the newly created resources. a) Is it just by convention that POST create request doesn't contain an URI of the newly created resource or POST create request actually can't contain the URI of the newly created resource? b) PUT has idempotent semantics and thus can be safely used for absolute updates ( ie we send entire state of the resource to the server ), but not also for relative updates ( ie we send just changes to the resource state ), since that would violate its semantics. But I assume it's still possible for PUT to send relative updates to the server, it's just that in that case the PUT update won't be idempotent? 2) I've read somewhere that we should "use POST to append a resource to a collection identified by a service-generated URI". a) What exactly does that mean? That if URIs for the resources were generated by a server ( thus the resources were created via POST ), then ALL subsequent resources should also be created via POST? Thus, in such situation no resource should be created via PUT? b) If my assumption under a) is correct, could you elaborate why we shouldn't create some resources via POST and some via PUT ( assuming server already contains a collection of resources created via POST )? REPLY: 1) Please correct me if I'm wrong, but from your post and from the link you've posted, it seems: a) The Request-URI in POST is interpreted by server as the URI of the service. Thus, it could just as easily be interpreted as an URI of a newly created resource, if server code was written to recognize Request-URI as such b) Similarly, PUT is able to send relative updates, it's just that service code is usually written such that it will complain if PUT updates are relative. 2) Usually, create has fallen into the POST camp, because of the idea of "appending to a collection." It's become the way to append a resource to a list of resources. I don't quite understand the reasoning behind the idea of "appending to a collection" and why this idea prefers POST for create. Namely, if we create 10 resources via PUT, then server will contain a collection of 10 resources and if we then create another resource, then server will append this resource to that collection ( which will now contain 11 resources )?! Uh, this is kinda confusing thank you

    Read the article

  • Is Life Better as a Manager? [closed]

    - by foreyez
    I've read similar posts to this, but I want to specifically target this question. This is mainly geared towards Developers that became Managers in their career. I want to know if you think being a Manager is a funner/better/more interesting life than a programmer's. (Or would you rather go back to being a programmer, if so why?) Thanks Note: Programmers that are not actively managing others please don't answer this question.

    Read the article

  • Are separate business objects needed when persistent data can be stored in a usable format?

    - by Kylotan
    I have a system where data is stored in a persistent store and read by a server application. Some of this data is only ever seen by the server, but some of it is passed through unaltered to clients. So, there is a big temptation to persist data - whether whole rows/documents or individual fields/sub-documents - in the exact form that the client can use (eg. JSON), as this removes various layers of boilerplate, whether in the form of procedural SQL, an ORM, or any proxy structure which exists just to hold the values before having to re-encode them into a client-suitable form. This form can usually be used on the server too, though business logic may have to live outside of the object, On the other hand, this approach ends up leaking implementation details everywhere. 9 times out of 10 I'm happy just to read a JSON structure out of the DB and send it to the client, but 1 in every 10 times I have to know the details of that implicit structure (and be able to refactor access to it if the stored data ever changes). And this makes me think that maybe I should be pulling this data into separate business objects, so that business logic doesn't have to change when the data schema does. (Though you could argue this just moves the problem rather than solves it.) There is a complicating factor in that our data schema is constantly changing rapidly, to the point where we dropped our previous ORM/RDBMS system in favour of MongoDB and an implicit schema which was much easier to work with. So far I've not decided whether the rapid schema changes make me wish for separate business objects (so that server-side calculations need less refactoring, since all changes are restricted to the persistence layer) or for no separate business objects (because every change to the schema requires the business objects to change to stay in sync, even if the new sub-object or field is never used on the server except to pass verbatim to a client). So my question is whether it is sensible to store objects in the form they are usually going to be used, or if it's better to copy them into intermediate business objects to insulate both sides from each other (even when that isn't strictly necessary)? And I'd like to hear from anybody else who has had experience of a similar situation, perhaps choosing to persist XML or JSON instead of having an explicit schema which has to be assembled into a client format each time.

    Read the article

  • Mexico Minimum Wage Changes - Payroll Patches Available

    - by LuciaC
    Mexico has published new Minimum Wage values effective November 27th 2012.  The following Payroll patches have been released to update the Minimum Wages:  Release  Patch Number  11i Patch 15919087  12.0.x Patch 15920839:R12.PAY.A  12.1.x Patch 15920839:R12.PAY.B Please note the following: the Minimum Wage values have been updated and they are effective from November 27th 2012 these patches are different from all other statutory updates (there are additional post installation steps), so please be sure to carefully read the entire patch readme before beginning to install these patches to ensure successful processing.

    Read the article

  • Oracle OpenWorld 2014 Preview: Don't-Miss Sessions, Hands-on Labs, and More

    - by Scott McNeil
    Check out all the latest Oracle Enterprise Manager 12c sessions at this year's Oracle OpenWorld. Organizers of the event, taking place in San Francisco from September 28 to October 2, expect heavy turnout at sessions, hands-on labs, and customer panels devoted to Oracle Enterprise Manager 12c. Find out who is participating and which sessions are most recommended by the Oracle Enterprise Manager team. Read More Stay Connected: Twitter | Facebook | YouTube | Linkedin | Newsletter Download the Oracle Enterprise Manager 12c Mobile app

    Read the article

  • Shows how to make your GridViews row-selectable and how to bind to a row's detail.

    How to make an ASP.NET GridView row-selectable and how to bind selected row detail to other controls. The key "trick" to learn is to be able to add an onclick javascript event to each row in the grid that will fire the intrinsic ASP.NET __doPostBack() method.  read moreBy Peter BrombergDid you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • JavaScript tip a day: More Debugging Tricks

    - by Sahil Malik
    SharePoint, WCF and Azure Trainings: more information Debugging is a pain. Debugging events on a web page is an especially bigger pain. This video will make that pain go away! This video will show you $ keywords, debugger statement, conditional breakpoints, monitoring events, global error handling etc. Make sure you check out the debugging video from yesterday too. Read full article ....

    Read the article

  • BI Beginner: Why to Use Excel 2013 Now

    - by John Paul Cook
    Most corporations and many individuals are slow to adopt new versions of Microsoft Office, particularly if the upgrade to the previous version was very recent. Excel 2013 is a special case and offers significant productivity enhancements. If you do business intelligence work or otherwise make your living with Excel, adding (notice I didn’t say upgrading to) Excel 2013 now makes a lot of sense. The Power View feature in Excel 2013 is a completely sufficient reason to add Excel 2013. It has to be enabled,...(read more)

    Read the article

  • 12.04 How to go back to command line only ubuntu

    - by Ne0
    I was attempting to install a command line only version of Ubuntu and followed the instructions here. I downloaded the alternate CD and went through the installation, but never got an "Install a command-line system." option. 12.04 is now up and running, though i don't need the desktop GUI. From what i have read i can set it back into text mode using grub, but what packages should i remove that i don't need any more?

    Read the article

  • ClearTrace for SQL Server 2012

    - by Bill Graziano
    I’ve updated the beta for ClearTrace that support SQL Server 2012.  This requires SQL Server 2012 to be installed on the computer where ClearTrace is running.  It will read traces from SQL Server 2008 R2, SQL Server 2008 and SQL Server 2005. It includes some minor improvements in performance and handling large SQL statements. It should also give better errors. If you do find any of those errors, please report them in the support forum.

    Read the article

  • SEO Tips - Avoiding 3 Common Mistakes

    Some of the most effective SEO tips are also some of the easiest to implement and are often overlooked. An effectively optimized site can increase the amount of free and targeted search engine traffic you receive. Read on to see 3 search engine optimization tips that are easy to follow and will result in more free traffic to your site.

    Read the article

  • Are We Losing a Standard (Edition) Data Recovery Technology?

    - by AllenMWhite
    One of the coolest technologies Microsoft released with SQL Server 2005 was Database Mirroring, which provided the ability to have a failover copy of a database on another SQL Server instance, and have the ability to automatically failover to that copy should a problem occur with the primary database. What was even cooler was that this new technology was available on Standard Edition! Mom and Pop shops could afford to implement a high availability solution without paying an extra tens of thousands...(read more)

    Read the article

  • Ubuntu Philosophy Question

    - by The-Ever-Kid
    I was just switching from Windows to Ubuntu and I started to read this. And there were quite a few things I did not understand one was : "OpenOffice decided not to have a learning curve" And "Firefox tries very hard to make sure pages written in 1995 look like they did in 1995. " And Finally "Windows isn't a poor man's Linux." In the final statement shouldn't the statement be the opposite.

    Read the article

  • Why wont SVN work on the Google Code Source? [closed]

    - by BluFire
    I installed SVN Toroise. I then proceeded to the desktop and right clicked and clicked on SVN checkout. I then entered under the URL of Repository "http://apple-crunch.googlecode.com/svn/trunk/ apple-crunch-read-only" but after a few seconds the checkout fails saying that the URL doesn't exist. I got the command line from http://code.google.com/p/apple-crunch/source/checkout I'm trying to get a direct copy instead of browsing through the source.

    Read the article

< Previous Page | 301 302 303 304 305 306 307 308 309 310 311 312  | Next Page >