Search Results

Search found 15051 results on 603 pages for 'meta programming'.

Page 154/603 | < Previous Page | 150 151 152 153 154 155 156 157 158 159 160 161  | Next Page >

  • Will learning ColdFusion help me advance my programming skills? [closed]

    - by chhantyal
    Currently I am working with a small web development company. We use jQuery on front-end, Coldfusion for back-end and MySQL as our database. We just started to use HTML5 and CSS3. This is my first internship and job. I know the basics of Python and want to add Django or Ruby on Rails on my skills set. In addition, I want to advance my programming skills with Machine Learning, Compilers, NoSQL and Unix Hacking. I also find front end web development pretty interesting. Should I focus on front-end and become skilled on HTML5/CSS3/Javascript? Or dive into back-end learning ColdFusion. I will probably leave the company after a year since I want to work with great product start-ups. And I live in India, where ColdFusion is not popular. Will learning ColdFusion help me become better programmer?

    Read the article

  • What is the best way to promote a programming blog?

    - by paul
    (The guys from 'Programmers' referred me here...) How do you promote your programming blog? I recently started http://blackforestcoder.blogspot.com/ to record my progress working with new technologies and ideas. The main aim being to provide a list of pitfalls and solutions and also to get feedback from readers. Since I set it up 10 days ago I have only had about 2-3 hits even though Google is supposed to be indexing it. How might I boost the hit rate?

    Read the article

  • Is there a difference between "self-plagiarizing" in programming vs doing so as a writer?

    - by makerofthings7
    I read this Gawker article about how a writer reused some of his older material for new assignments for different magazines. Is there any similar ethical (societal?) dilemma when doing the same thing in the realm of Programming? Does reusing a shared library you've accumulated over the years amount to self-plagarizm? What I'm getting at is that it seems that the creative world of software development isn't as stringent regarding self-plagarism as say journalism or blogging. In fact on one of my interviews at GS I was asked what kind of libraries I've developed over the years, implying that me getting the job would entail co-licensing helpful portions of code to that company. Are there any cases where although it's legal to self-plagarize, it would be frowned upon in the software world?

    Read the article

  • How Can Someone With Minimum Programming Experience Publish an App of an Existing Program not found in the Software Centre?

    - by Brenton Horne
    There's a free mathematics software that I would like to see in Ubuntu Software Centre, called, 'Sage.' I have minimal experience in programming, so I'm basically here asking if someone could either tell me how to accomplish this without said instructions being completely over my head or if someone could do it for me. For further information here's the sage homepage: http://sagemath.org/ and the .tar.lzma file for its installation is ~500MB and it takes around 2.2 GB on the HDD when installed, I found at least on 32 bit i386 Ubuntu 12.04 LTS. Any help will be greatly appreciated!

    Read the article

  • How can I get started programming OpenGL on Mac OS X?

    - by Michael Stum
    I'm trying to start OpenGL programming on a Mac, which brings me into unknown territory on a lot of things. During the day, I'm a Web Developer, working in C# and before that in PHP and Delphi, all on Windows. During the night, I try to pick up Mac/OpenGL skills, but everything is so different. I've been trying to look for some books, but the OpenGL books are usually for iOS (tons of them out there) and the Mac Books usually cover "normal" application Development. I want to start simple with Pong, Tetris and Wolfenstein. I see that there are a bunch of different OpenGL Versions out there. I know about OpenGL ES 1&2, but I don't know about the "big" OpenGL Versions - which ones are commonly supported on 10.6 and 10.7 on current (2010/2011) Macs? Are there any up to date (XCode 4) books or tutorials? I don't want to use a premade Engine like Unity yet - again, I know next to nothing about any Mac development.

    Read the article

  • What is the best way to promote a programming blog?

    - by paul
    (The guys from 'Programmers' referred me here...) How do you promote your programming blog? I recently started http://blackforestcoder.blogspot.com/ to record my progress working with new technologies and ideas. The main aim being to provide a list of pitfalls and solutions and also to get feedback from readers. Since I set it up 10 days ago I have only had about 2-3 hits even though Google is supposed to be indexing it. How might I boost the hit rate?

    Read the article

  • Lean/Kanban *Inside* Software (i.e. WIP-Limits, Reducing Queues and Pull as Programming Techniques)

    - by Christoph
    Thinking about Kanban, I realized that the queuing-theory behind the SW-development-methodology obviously also applies to concurrent software. Now I'm looking for whether this kind of thinking is explicitly applied in some area. A simple example: We usually want to limit the number of threads to avoid cache-thrashing (WIP-Limits). In the paper about the disruptor pattern[1], one statement that I found interesting was that producer/consumers are rarely balanced so when using queues, either consumers wait (queues are empty), or producers produce more than is consumed, resulting in either a full capacity-constrained queue or an unconstrained one blowing up and eating away memory. Both, in lean-speak, is waste, and increases lead-time. Does anybody have examples of WIP-Limits, reducing/eliminating queues, pull or single piece flow being applied in programming? http://disruptor.googlecode.com/files/Disruptor-1.0.pdf

    Read the article

  • Is OO-programming really as important as hiring companies place it?

    - by ale
    I am just finishing my masters degree (in computing) and applying for jobs.. I've noticed many companies specifically ask for an understanding of object orientation. Popular interview questions are about inheritance, polymorphism, accessors etc. Is OO really that crucial? I even had an interview for a programming job in C and half the interview was OO. In the real world, developing real applications, is object orientation nearly always used? Are key features like polymorphism used A LOT? I think my question comes from one of my weaknesses.. although I know about OO.. I don't seem to be able to incorporate it a great deal into my programs. I would be really interested to get peoples' thoughts on this!

    Read the article

  • Is Ruby on Rails' Active Record an example of Aspect-Oriented Programming?

    - by B Seven
    From Clean Code, about Cross-Cutting Concerns: Note that concerns like persistence tend to cut across the natural object boundaries of a domain. You want to persist all your objects using generally the same strategy, for example, using a particular DBMS... Is Active Record an example of aspect-oriented programming? In AOP, modular constructs called aspects specify which points in the system should have their behavior modified in some consistent way to support a particular concern. This specification is done using a succinct declarative or programmatic mechanism. If Active Record is an example of AOP, what is the "aspect"? Is it the class declaration that inherits from Active Record? class Foo < ActiveRecord::Base

    Read the article

  • Which specific programming activities do women, on average, perform better than men? [closed]

    - by blueberryfields
    Following a recent discussion with female associates in hiring positions for software development/engineering positions, I found out that this kind of information would be incredibly useful to helping make sure that the workforce shows a gender balance. So I went looking. I've found various literature speaking about risk-taking behaviour and patterns, and other statistical differences between men and women when it comes to work performance. See for example this article related to hedge fund management. I have yet to see any such comparison in the computing field. To restate the question: Which specific programming activities do women, on average, perform better than men? Please back up your answers with specific details, preferably by linking to relevant research or, failing that, explaining what you're basing the information on.

    Read the article

  • Are there similarities between operating system kernels and programming language kernels?

    - by rahmu
    I know very little about Smalltalk but I noticed that there's a frequent mention of the "kernel". Dan Ingalls prime maintainer of several implementations of Smalltalk also worked on a Javascript environment called "Lively Kernel" and in Peter Siebel's book he kept mentionning the "kernel". I cannot help but think that it is no coincidence that the creators of Smalltalk used the name of a (central) part of operating systems to refer to a particular component of their language. Was it because Smalltalk intended to act as an operating system? Was it because theory behind programming languages and operating systems have a lot in common? What is the reason behind the common appelation of the two components?

    Read the article

  • Why are cryptic short identifiers still so common in low-level programming?

    - by romkyns
    There used to be very good reasons for keeping instruction / register names short. Those reasons no longer apply, but short cryptic names are still very common in low-level programming. Why is this? Is it just because old habits are hard to break, or are there better reasons? For example: Atmel ATMEGA32U2 (2010?): TIFR1 (instead of TimerCounter1InterruptFlag), ICR1H (instead of InputCapture1High), DDRB (instead of DataDirectionPortB), etc. .NET CLR instruction set (2002): bge.s (instead of branch-if-greater.signed), etc. Aren't the longer, non-cryptic names easier to work with?

    Read the article

  • Best options for freelance or part-time programming? [closed]

    - by Andrew
    I apologize in advance if this is an inappropriate question for this SE. A few years back I was all set to study computer science and get a job programming, but went a totally different route and went into healthcare. I currently work as a paramedic on a rotating 24/48 schedule, so I have two days off for every day I work, and a decent bit of downtime on the days that I do work. I've been looking at ways to earn some extra money with all that spare time, and was wondering if it'd be worth the effort to try and find a part-time/freelance gig. I know HTML/CSS, PHP, and I'm pretty familiar with Python and Ruby (and Rails). Anyways, was hoping that someone could point me in the right direction as what "skill set" would give me the best chance to be able to land a part-time/freelance gig. I realize this is a rather open-ended question but any direction is appreciated.

    Read the article

  • What are startups expecting when they ask you to solve a programming challenge before interviewing? [closed]

    - by Swapnil Tailor
    I have applied to couple of startups and most of them are asking to solve programming challenge before they start on the interviewing candidate. I have submitted couple of the solutions and all the time getting rejected in the initial screening. Now what I think is, they will see my coding style, algorithm and OOD concepts that I have used to solve the problem. Can you guys input more on it as what other details are taken into consideration and how can I improve my coding for getting selected. By the way, I did all my coding in either Java/Perl. EDIT I feel the biggest reason for rejection is code didn't work for couple of use cases.

    Read the article

  • When to learn the command line version of a programming tool ?

    - by explorest
    Almost every programming tool has a command line version; many of which also have a gui version. It takes a lot of time and memorization effort to learn the different commands and various options/switches om the command line version. So I have a couple of questions (which are not necessarily mutually exclusive): 1) When would you bother learning/memorizing the commands in the command line version of a tool which also comes in a gui version ? 2) What tools should I learn the command line version of ? .... compilers ? version control system ? etc, etc

    Read the article

  • Anti-depressant and programming: does it reduce your thinking and focusing abilities ?

    - by user12358
    I wanted to ask your opinion on anti-depressants, since I took them daily for 3 years now, but I can't be sure if I'm less perfomant with them or without, since I never withdrawed. I'm still at school at the age of 25, still having some motivation problems (for example I can't get used to do something at school if I don't think it will teach me something), but I'm quite motivated to work in the video-game field, since I have some personnal projects in mind. I know C++ programming etc, I'm still learning techniques, but do you think I should try more to do my project instead of just following the work I'm assigned to ? Have you had experience with depression or anti-depressants ? How did it affect your work ? Do you think that being depressed or half-depressed can improve creativity ? Do you feel it has affected the way you end up writing your algorithms ? EDITED

    Read the article

  • When to learn the command line version of a programming tool?

    - by explorest
    Almost every programming tool has a command line version; many of which also have a gui version. It takes a lot of time and memorization effort to learn the different commands and various options/switches of the command line version. So I have a couple of questions (which are not necessarily mutually exclusive): 1) When would you bother learning/memorizing the commands in the command line version of a tool which also comes in a gui version ? 2) What tools should I learn the command line version of ? .... compilers ? version control system ? etc, etc

    Read the article

  • is it possible to get a programming job in the bay area?

    - by user475119
    Is it my perception or is it reality that it is harder to get software development work in Silicon Valley, than in other population centers? My theory is that some of the best and brightest (and youngest) are all here. All working for the top tech companies. And, little ol' me, over 40, math degree (not at Standford or Berkeley), ok at algorithms, and with many skills with many programming tools and concepts, but not published, and not top my class has a bit of a tough time competing. (I'm a Python programmer, by the way.) Thanks

    Read the article

  • What would be topic for research in on edge of multiple processors / computers programming?

    - by Kabumbus
    I mean what is not already there? What can be developed in fiew month and give a breakthrue/ start a new leap in science of f multiple computers programming? What i see is already there MPI/ Bit torrent/Jabber protocols / APIs / servers for messaging LAN / wire and other infrastractural cabels for connecting Boost and analogs on evry OS in most languages for multithreading there are lots of CUDA like on computer frameworks for fast calculating on computers GPUs What I personally do not see out there is a crossplatform framework for multiple processes interaction. Meaning one that would allow easy creation of multyple processes running in paralell inside one hoster app on one machine. In level not harder than needed for threads creation (so no seprate server apps - just one lib doing it all) Is there ny such lib and what can you propose for research topic?

    Read the article

  • Is there such a thing these days as programming in the small?

    - by WeNeedAnswers
    With all the programming languages that are out there, what exactly does it mean to program in the small and is it still possible, without the possibility of re-purposing to the large. The original article which mentions in the small was dated to 1975 and referred to scripting languages (as glue languages). Maybe I am missing the point, but any language that you can built components of code out of, I would regard to being able to handle "in the large". Is there a confusion on what Objects are and do they really figure as being mandatory to being able to handle "the large". Many have argued that this is the true meaning of "In the large" and that the concepts of objects are best fit for the job.

    Read the article

  • Is "convention over configuration" not violating basic programming principles?

    - by Geerten
    I was looking at the WPF MVVM framework Caliburn.Micro and read that a lot of standard things are based on naming conventions. For example, automatic binding of properties in the View to properties in the ViewModel. Although this seems to be convenient (removes some boilerplate code), my first instinct reaction is that it isn't completely obvious to a new programmer that will read this code. In other words, the functionality of the application is not completely explained by its own code, but also by the documentation of the framework. EDIT: So this approach is called convention over configuration. Since I could not find any questions concerning this, I altered my question: My question is: Is convention over configuration a correct way of simplifying things, or is it violating some programming principles (and if so, which ones)?

    Read the article

  • Is there anything in .NET that allows me to define a grammar and generate a programming language?

    - by user1525474
    I have a course in which the proffesor has asked us to create a DSL for a our final project. He presented us in the first courses xText with Eclipse. This being a new course, I am still a bit fuzzy on what Domain Specific Languages means. This is my current understanding: a domain specific language is a language that is created for specific problems in software development. Examples of DSL's are PHP, SQL, JavaScript and on the opposite are languages like Java , C# , C++ , Ruby etc. Please feel free to correct me if I am wrong. What I would like to know: is there is any tool for .NET/Visual Studio that is similar to Xtext, that allows me to define a grammar and be allowed to generate a programming language based on that with an activity diagram?

    Read the article

  • Starting Game Programming in C++, where to begin for optimal cross-platform development?

    - by Qasim
    I wish to begin learning game development in C++. I have knowledge in syntax and the console using Visual C++ 2010, however I want to begin programming. My only problem is that I want to be able to support many platforms including Windows, Mac OS, and Linux. As I get better, I also hope I can submit games to XBLIG (or XBLA) and Steam, and even the Google Play Store (But that would take some porting and major redesign of controls). I have looked into c++ and SDL but I have no idea how to get started. I have Visual C++ 2010 installed and I hope I can still use that as other C++ IDEs are quit outdated. Because I want to stick with Visual C++ and SDL (unless there is a better library than SDL), there is not much room for debate but rather how to create projects to support cross-platform development.

    Read the article

  • What language should an 11-year old start with to learn game programming?

    - by emsr
    I have a 11-year old son who wants to do game programming. I've started him on C++ (C++11) and he's learned iostreams, looping, functions, logic and flow control. I'm using the standard library and no memory management at all. But I would like to ask: What language would you suggest for a pre-teen (Python, ...)? What books would you suggest? We looked at one book that was just for console ASCII games. I liked the C++ that it taught but I think he'll get bored without some graphics at some point.

    Read the article

  • How can I effectively use a netbook and a desktop computer together for programming?

    - by Mana
    Currently, in my workspace, I have a netbook sitting off to the side gathering dust while I write code on my desktop. As a result, the only use my netbook gets coding-wise is when I'm writing up a quick Python script to model a given problem or concept in class; I never use it at home for coding, or for anything at all, as it is all possible and faster on my (much more powerful) desktop. I feel like this is wrong and that I should be making better use of my netbook. What effective uses have you found for a netbook and a desktop together when programming (or for software development in general)? What are the merits of this practice?

    Read the article

< Previous Page | 150 151 152 153 154 155 156 157 158 159 160 161  | Next Page >