Search Results

Search found 18140 results on 726 pages for 'language learning'.

Page 52/726 | < Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >

  • Moving from VB.NET to C#

    - by w0051977
    I have worked with the VB.NET programming language for the last five years. I want to move to C# as I believe skills are more valued plus it is more similar to other object oriented languages like Java. I was offered a job today working primarily with C#. I explained at the interview that I am a VB.NET Developer and I did the test in VB.NET (though they would of preferred C#). If I decide to accept the position then I will be starting at at the top end of the salary bracket (only very slightly more than I earn now). I will have to help more junior staff in the future who probably have more experience using C# than I do (1-2 years). I used Java and C++ at university. I want to move towards C# in the future as I believe C# skills are more valued based on job advertisements I have seen recently. Has anyone else done this and did it work? i.e. move to a new organisation as a C# Developer at quite a senior level with experience primarily using VB.NET.

    Read the article

  • How to progress far enough in a year [on hold]

    - by xCasper
    So I came to a realization the other day, I graduate in a year. I will have completed my four year degree in about two and a half (I went to a school that goes full time year round). Anyway, I want to get a job, as a programmer, when I graduate. The problem is, I feel like I am no where near ready. In the last year I cant say I have made any spectacular projects. I know that my advanced programming courses are coming up, but by the time I take them, I will be graduating in 6 months. Not nearly enough time to really take what I learn, apply it, and create something to show for myself. I want to push myself ahead of the game; mainly because my major is Computer Information Systems, so the focus is not programming. In fact, I only get, 4 programming classes. Before anyone says anything, CIS is the closest to a programming major at my school that I am able to do. So the questions come down to this: 1) What can I do to really step up the speed at which I progress (on my own) 2) Should I be aiming for a certain amount of projects in my "Portfolio." 2a) Should they be big projects? P.S: The language we have used in school in c++, I do take a Java class in the spring, and .net over Summer; if any of that matters for anything.

    Read the article

  • Looking for some advice on the next steps to take [closed]

    - by mopsyd
    I am looking for some advice on the next step to take in development of my programming skills. I was directed here when asking this question on Stack Overflow. What I know already Have a solid grasp of xhtml, xml, php, javascript, MySQL, actionscript. Have a working knowledge of vb, and have a slight grasp of java from tinkering with a minecraft server. Some brief exposure to the Unreal Engine in college. Some skills with sql server, ms sql, office integration, etc. Also some knowledge of Asterix and PBX/VOIP. Been coding off and on since the age of 8 but I have no computer science education aside from what I have taught myself or learned from work/freelance. I work in OSX mostly, but can use/troubleshoot windows and ubuntu fluently also. Decent with both UNIX and DOS CLI. What I'm considering I'm looking to learn a scripting language to build web apps, help streamline my home server that I am building and run shell scripts. Being able to help code games later is a big plus. My Question Between java, ruby, perl, and python, which would be the best investment of my time considering what I already know and what direction I would like to take my skillset? What are good resources for your suggested direction? Thanks in advance.

    Read the article

  • Noob Rails ? about learning Rails

    - by user271916
    Hi All I have been programming for a while and for the past 3 or 4 months have been learning ruby. I am not an expert by any means but I believe I have the basics down. I decided to start learning RoR and bought the "Agile Web Development with Rails 3rd Edition" and have been dutifully going through the chapters one by one. Currently I am in chapter 8 and have had no problems so far. My question is I know I have learned several things so far and I know that I am starting to get a sense of the Rails framework I have this fear that I am just not learning as much as I should. Some things I get and understand the interconnections while I feel on other things I am just going through the motions and don't fully comprehend the total interconnectivity. Now, there is still a large amount of the book for me to complete. I guess I am just wondering if I complete this book what should I expect to be able to accomplish on my own and what should be my next steps. Thanks

    Read the article

  • How should I go about learning Python?

    - by user345690
    I am currently learning PHP and want to learn about OOP. *1.*I know Python is a well-organized and is all OOP, so would learning Python be a wise choose to learn OOP? The thing is I am more towards web development then just general programming, and I know Python is just a general purpose language, but there is Django. *2*So how should I go about learning Python if I am lending towards web development? Is there any good books/websites that help me learn Python for web development? 3. Is there any free webhosting companies that allow Python? I never used Python before, only PHP, and not sure how it works? Can I run Python on a localhost?

    Read the article

  • Problem with document.location.href

    - by novellino
    Hello, I am new to Javascript and Web development and I have a question regarding the document.location.href. I am using a cookie for storing the language the user prefers and then load the english or the swedish version depending on the language. The default language in the beginning is the same as the browser's language, and my index.jsp is the swedish one. The first time everything works fine. The problem is when the cookie exists already. The basic code is: if (language!=null && language!=""){ if (language=="en-US" || language=="en-us") document.location.href = "en/index.jsp"; } else{ //Explorer if (navigator.userLanguage) language = navigator.userLanguage; //other browsers else language = (navigator.language) ? navigator.language : navigator.userLanguage; if (language!=null && language!=""){ setCookie('language', language, 365, '/', 'onCheck'); if (language=="en-US" || language=="en-us") document.location.href = "en/index.jsp"; else if(language=="sv") document.location.href="index.jsp"; } } When the cookie exists we enter the first "if", and there, if the language is swedish it opens the default blabla/index.jsp page. When the language is set to engish it should open the blabla/en/index.jsp but instead it opens the blabla/en/en/index.jsp which of course is wrong. Does anyone know what I am doing wrong?? Thanks

    Read the article

  • Best Advice Ever: Learn By Helping Others

    - by Argenis
    I remember when back in 2001 my friend and former SQL Server MVP Carlos Eduardo Rojas was busy earning his MVP street-cred in the NNTP forums, aka Newsgroups. I always thought he was playing the Sheriff trying to put some order in a Wild Wild West town by trying to understand what these people were asking. He spent a lot of time doing this stuff – and I thought it was just plain crazy. After all, he was doing it for free. What was he gaining from all of that work? It was not until the advent of Twitter and #SQLHelp that I realized the real gain behind helping others. Forget about the glory and the laurels of others thanking you (and thinking you’re the best thing ever – ha!), or whatever award with whatever three letter acronym might be given to you. It’s about what you learn in the process of helping others. See, when you teach something, it’s usually at a fixed date and time, and on a specific topic. But helping others with their issues or general questions is something that goes on 24x7, on whatever topic under the sun. Just go look at sites like DBA.StackExchange.com, or the SQLServerCentral forums. It’s questions coming in literally non-stop from all corners or the world. And yet a lot of people are willing to help you, regardless of who you are, where you come from, or what time of day it is. And in my case, this process of helping others usually leads to me learning something new. Especially in those cases where the question isn’t really something I’m good at. The delicate part comes when you’re ready to give an answer, but you’re not sure. Often times I’ll try to validate with Internet searches and what have you. Often times I’ll throw in a question mark at the end of the answer, so as not to look authoritative, but rather suggestive. But as time passes by, you get more and more comfortable with that topic. And that’s the real gain.  I have done this for many years now on #SQLHelp, which is my preferred vehicle for providing assistance. I cannot tell you how much I’ve learned from it. By helping others, by watching others help. It’s all knowledge and experience you gain…and you might not be getting all that in your day job today. Such thing, my dear reader, is invaluable. It’s what will differentiate yours amongst a pack of resumes. It’s what will get you places. Take it from me - a guy who, like you, knew nothing about SQL Server.

    Read the article

  • How do you take into account usability and user requirements for your application?

    - by voroninp
    Our team supports BackOffice application: a mix of WinForm and WPF windows. (about 80 including dialogs). Really a kind of a Swiss Army Knife. It is used by developers, tech writers, security developers, testers. The requirements for new features come quite often and sometimes we play Wizard of Oz to decide which GUI our users like the most. And it usually happens (I admit it can be just my subjective interpretation of the reality) that one tiny detail giving the flavor of good usability to our app requires a lot of time. This time is being spent on 'fighting' with GUI framework making it act like we need. And it very difficult to make estimations for this type of tasks (at least for me and most members of our team). Scrum poker is not a help either. Management often considers this usability perfectionism to be a waste of time. On the other hand an accumulated affect of features where each has some little usability flaw frustrates users. But the same users want frequent releases and instant bug fixes. Hence, no way to get the positive feedback: there is always somebody who is snuffy. I constantly feel myself as competing with ourselves: more features - more bugs/tasks/architecture. We are trying to outrun the cart we are pushing. New technologies arrive and some of them can potentially help to improve the design or decrease task implementation time but these technologies require learning, prototyping and so on. Well, that was a story. And now is the question: How do you balance between time pressure, product quality, users and management satisfaction? When and how do you decide to leave the problem with not a perfect but to some extent acceptable solution, how often do you make these decisions? How do you do with your own satisfaction? What are your priorities? P.S. Please keep in mind, we are a BackOffice team, we have neither dedicated technical writer nor GUI designer. The tester have joined us recently. We've much work to do and much freedom concerning 'how'. I like it because it fosters creativity but I don't want to become too nerdy perfectionist.

    Read the article

  • Windows 7 keeps changing my language(keyboard) settings

    - by Hans
    I am using Windows 7 Ultimate on a Windows network. My default settings is Danish locale, Danish keyboardlayout, but English system language. I only have Danish keyboard layout installed (in Text Services and Input Languages). However, at the login screen I can choose Danish and English, but whatever I choose Windows 7 changes system language (and keyboard layout!) to English (United States). Since this is a per-App setting I must change the language setting (language bar in the taskbar) Where can this be changed system wide and permament?

    Read the article

  • Heap corruption error after language pack installation for Visual Studio 2012

    - by Lyndon
    I have installed the german version of Visual Studio 2012 Premium on my german windows machine and installed the english language pack vor Visual Studio 2012 Premium and it works great but after I installed the german language pack I get the heap corruption error 0xc0000374. The faulty module is ntdll.dll, version: 6.3.9600.16408 Only restoring Windows resolves this issue. Edit: This error also occurs when changing the displayed language and I was able to observe this behavior only after updating from Windows 8 to Windows 8.1 and updating from DevExpress 12.1 to DevExpress 13.1. Not only that, but the error does not occure immediately after installing a language, sometimes I can start debugging my program as usual and then after three to five times or so, the error occurs. Is there another solution than restoring Windows?

    Read the article

  • language input is changed by itself, not only by keystroke of alt+shift

    - by Berry Tsakala
    I'm using 2 and sometimes more input methods in windows I use ALT + left Shift for switching between languages. Sometimes, every once in a while, another action(s) trigger the language switch. It happens in the same application, (i.e. the input language changes for the same app without my request) or, while switching tasks with Alt-Tab, which causes language input to change in an arbitrary application's context. I realized that some of these actions could be the use of the scroll wheel button. It's super annoying How do I disable any other language switching, and stay only with alt+shift? Why does this happen? It happened to my in the past in these and more occasions: middle mouse click pressing Windows key pression Alt Tab (unknown)

    Read the article

  • Firefox: fast dictionary language switching

    - by Lo'oris
    I have two language dictionaries installed on Firefox 3.6, and I would like to be able to switch really fast between them, using the keyboard. At the moment the only way I know I can switch is right clicking in a text input field, go into Language, and then click the language. I would instead to be able to switch between those two just hitting two keys at most, if possible just one (something like F13). Searching for addons I've found tons of extensions somewhat related but that don't actually do what I want.

    Read the article

  • Why is C++ often the first language taught in college?

    - by Casey Patton
    My school starts the computer science curriculum with C++ programming courses, meaning this is the first language that many of the students learn. I've seen that many people dislike C++, and I've read a variety of reasons why. It almost seems to be popular opinion that C++ isn't a very good language. I get the impression it's not very liked based on some questions on StackExchange as well as posts such as: http://damienkatz.net/2004/08/why-c-sucks.html http://blogs.kde.org/node/2298 http://blogs.cio.com/esther_schindler/linus_torvalds_why_c_sucks http://www.dacris.com/blog/2010/02/16/why-c-sucks-part-2/ etc. (Note: It is not my opinion that C++ is a bad language. In fact, it's the main language I use. However, the internet as well as some professors have given me the impression that it's not a very widely liked language. In fact, one of my professor constantly rags on C++, yet it's still the starting language at my college!) With that in mind, why is this the first language taught at many schools? What are the reasons for starting a programming curriculum with C++? Note: This question is similar to "Is C++ suitable as a first language", but is a little different since I'm not interested in whether it's suitable, but why it's been chosen.

    Read the article

  • Managed Languages vs Compiled Language difference?

    - by l46kok
    I get confused when people try to make a distinction between compiled languages and managed languages. From experience, I understand that most consider compiled languages to be C,C++ while managed languages are Java,C# (There are obviously more, but these are just few examples). But what exactly is the core difference between the two types of languages? My understanding is that any program, regardless of what language you use is essentially "compiled" into a low-level machine code which is then interpreted, so does that kinda make managed languages a subset of compiled languages (That is, all managed languages are compiled languages but not the other way around)?

    Read the article

  • How can we improve overall Programmer Education & Training?

    - by crosenblum
    Last week, I was just viewing this amazing interview by Kevin Rose of Phillip Rosedale, of Second Life. And they had an amazing discussion about how to find, hire and identify good programmer's, and how hard it is to find good ones. Which has lead me to really think about the way we programmer's learn, are taught. For a majority of us, myself included, we are self-taught. Which is great about being a programmer, anyone can learn and develop skills. But this also means, that there is no real standards of what a good programmer is/are, and what kind of environment's encourage the growth of programming skills. This isn't so much a question, but just a desire in me, to see how we can change the culture of programming, and the manager's of programming, so that education and self-improvement is encouraged. There are a lot of avenue's for continued education, youtube videos, books, conferences, but because of the experiental nature of what we do, it isn't always clear what's important to learn and to master. Let's look at the The Joel 12 Steps. The Joel Test Do you use source control? Can you make a build in one step? Do you make daily builds? Do you have a bug database? Do you fix bugs before writing new code? Do you have an up-to-date schedule? Do you have a spec? Do programmers have quiet working conditions? Do you use the best tools money can buy? Do you have testers? Do new candidates write code during their interview? Do you do hallway usability testing? I think all of these have important value, but because of something I call the Experiential Gap, if a programmer or manager has never experienced any of the negative consequences for not having done items on the list, they will never see the need to do any of them. The Experiental Gap, is my basic theory, that each of us has different jobs and different experiences. So for some of us, that have always worked with dozens of programmer's, source control is a must have. But for people who have always been the only programmer, they can not imagine the need for source control. And it's because of this major flaw in how we learn, that we evaluate people by what best practices they do or not do, and the reason for either can start a flame war. We always evaluate people in our field by what they do, and think "Oh if this guy/gal isn't doing xyz best practice, he/she can't be a good programmer, so let's not waste time or energy talking to them." This is exactly why we have so many programming flame wars, that it becomes, because of the Experiental Gap, we can't imagine people not having made the decisions that we have had to made. So this has lead me to think, that we totally need to rethink how we train, educate and manage programmer's. For example, what percentage of you have had encouragement by your manager's to go to conferences, and even have them pay for it? For me, and a lot of people, this is extremely rare, a lot of us would love to go to conferences, to learn more, but the money ain't there to do that. So the point of this question is really to spark a lot of how can we train, learn and manage better? How can we create a new culture of learning that doesn't insult people for not having the same job experiences. Yes we all have jobs and work to do, but our ability to do our jobs well, depends on our desire, interest and support in improving our mastery of our skills. Right now, I see our culture being rather disorganized, we support the elite, but those tons of us that want to get better, just don't have enough support to learn and improve ourselves. I mean, do we as an industry, want to be perceived as just replaceable cogs? Thank you...

    Read the article

  • Design patterns and multiple programming language

    - by Eduard Florinescu
    I am referring here to the design patterns found in the GOF book. First how I see it, there are a few peculiarities to design pattern and knowing multiple language knowledge, for example in Java you really need a singleton but in Python you can do without it you write a module, I saw somewhere a wiki trying to write all GOF patterns for JavaScript and the entries where empty, I guess because it might be a daunting task. If there is someone who is using design patterns and is programming in multiple programming languages supporting the OOP paradigm and can give me a hint on how should I approach design patterns that might help me in all languages I use(Java, JavaScript, Python, Ruby): Can I write good application without knowing exactly the GOF design patterns or I might need some of them which might be crucial and if yes which one, are they alternatives to GOF for specific languages, and should a programmer or a team make its own design patterns set?

    Read the article

  • Mixed Emotions: Humans React to Natural Language Computer

    - by Applications User Experience
    There was a big event in Silicon Valley on Tuesday, November 15. Watson, the natural language computer developed at IBM Watson Research Center in Yorktown Heights, New York, and its inventor and principal research investigator, David Ferrucci, were guests at the Computer History Museum in Mountain View, California for another round of the television game Jeopardy. You may have read about or watched on YouTube how Watson beat Ken Jennings and Brad Rutter, two top Jeopardy competitors, last February. This time, Watson swept the floor with two Silicon Valley high-achievers, one a venture capitalist with a background  in math, computer engineering, and physics, and the other a technology and finance writer well-versed in all aspects of culture and humanities. Watson is the product of the DeepQA research project, which attempts to create an artificially intelligent computing system through advances in natural language processing (NLP), among other technologies. NLP is a computing strategy that seeks to provide answers by processing large amounts of unstructured data contained in multiple large domains of human knowledge. There are several ways to perform NLP, but one way to start is by recognizing key words, then processing  contextual  cues associated with the keyword concepts so that you get many more “smart” (that is, human-like) deductions,  rather than a series of “dumb” matches.  Jeopardy questions often require more than key word matching to get the correct answer; typically several pieces of information put together, often from vastly different categories, to come up with a satisfactory word string solution that can be rephrased as a question.  Smarter than your average search engine, but is it as smart as a human? Watson was especially fast at descrambling mixed-up state capital names, and recalling and pairing movie titles where one started and the other ended in the same word (e.g., Billion Dollar Baby Boom, where both titles used the word Baby). David said they had basically removed the variable of how fast Watson hit the buzzer compared to human contestants, but frustration frequently appeared on the faces of the contestants beaten to the punch by Watson. David explained that top Jeopardy winners like Jennings achieved their success with a similar strategy, timing their buzz to the end of the reading of the clue,  and “running the board”, being first to respond on about 60% of the clues.  Similar results for Watson. It made sense that Watson would be good at the technical and scientific stuff, so I figured the venture capitalist was toast. But I thought for sure Watson would lose to the writer in categories such as pop culture, wines and foods, and other humanities. Surprisingly, it held its own. I was amazed it could recognize a word definition of a syllogism in the category of philosophy. So what was the audience reaction to all of this? We started out expecting our formidable human contestants to easily run some of their categories; however, they started off on the wrong foot with the state capitals which Watson could unscramble so efficiently. By the end of the first round, contestants and the audience were feeling a little bit, well, …. deflated. Watson was winning by about $13,000, and the humans had gone into negative dollars. The IBM host said he was going to “slow Watson down a bit,” and the humans came back with respectable scores in Double Jeopardy. This was partially thanks to a very sympathetic audience (and host, also a human) providing “group-think” on many questions, especially baseball ‘s most valuable players, which by the way, couldn’t have been hard because even I knew them.  Yes, that’s right, the humans cheated. Since Watson could speak but not hear us (it didn’t have speech recognition capability), it was probably unaware of this. In Final Jeopardy, the single question had to do with law. I was sure Watson would blow this one, but all contestants were able to answer correctly about a copyright law. In a career devoted to making computers more helpful to people, I think I may have seen how a computer can do too much. I’m not sure I’d want to work side-by-side with a Watson doing my job. Certainly listening and empathy are important traits we humans still have over Watson.  While there was great enthusiasm in the packed room of computer scientists and their friends for this standing-room-only show, I think it made several of us uneasy (especially the poor human contestants whose egos were soundly bashed in the first round). This computer system, by the way , only took 4 years to program. David Ferrucci mentioned several practical uses for Watson, including medical diagnoses and legal strategies. Are you “the expert” in your job? Imagine NLP computing on an Oracle database.   This may be the user interface of the future to enable users to better process big data. How do you think you’d like it? Postscript: There were three little boys sitting in front of me in the very first row. They looked, how shall I say it, … unimpressed!

    Read the article

  • ibus symbol disappears for Japanese language selection

    - by Christian Becker-Asano
    Similar to this post No ibus icon in Kubuntu, I have trouble with the ibus icon in Ubuntu 12.04. Each time an update is installed, the language selector for Japanese disappears from the top panel. I need to uninstall and install Japanese again, then reboot, to make the symbol appear again. In this thread No iBus icon in Kubuntu 12.04 the suggestion was to install the Japanese Version of Ubuntu, but is it really true that one has to stick to a special version of Ubuntu to get this problem solved? If so, how can I transfer the settings easily from the current version to the Japanese one?

    Read the article

  • Generic and type safe I/O model in any language

    - by Eduardo León
    I am looking for an I/O model, in any programming language, that is generic and type safe. By genericity, I mean there should not be separate functions for performing the same operations on different devices (read_file, read_socket, read_terminal). Instead, a single read operation works on all read-able devices, a single write operation works on all write-able devices, and so on. By type safety, I mean operations that do not make sense should not even be expressible in first place. Using the read operation on a non-read-able device ought to cause a type error at compile time, similarly for using the write operation on a non-write-able device, and so on. Is there any generic and type safe I/O model?

    Read the article

  • Stack vs queue -based programming language efficiency [closed]

    - by Core Xii
    Suppose there are two programming languages; one where the only form of storage is one (preferred) or two (may be required for Turing-completeness) stacks, and another where the only form of storage is a single queue, with appropriate instructions in each to manipulate their respective storage to achieve Turing-completeness. Which one can more efficiently encode complex algorithms? Such that most given algorithms take less code to implement, less time to compute and less memory to do so. Also, how do they compare to a language with a traditional array (or unbounded tape, if you will) as storage?

    Read the article

  • Non-language-specific interview questions for a senior web developer

    - by Songo
    I came across a job posting for a senior web developer position. The posting said that the development will be done using Ruby on Rails, but no prior knowledge is required. I confirmed with a contact in that company that a PHP web developer can apply for it or even an ASP.Net developer. I also confirmed that the interview won't contain any questions specific to PHP or Ruby on Rails. Can anyone please provide a good list of questions for a senior web developer that isn't specific to a certain language? Note This question isn't a duplicate for similar posts asking for questions relating to PHP, .Net or Ruby. Also, I'm not looking for topics to learn as a web developer, but rather interesting questions for a technical interview given the former conditions.

    Read the article

  • Does syntax really matter in a programming language?

    - by Saif al Harthi
    One of my professors says "the syntax is the UI of a programming language", languages like Ruby have great readability and it's growing, but we see a lot of programmers productive with C\C++, so as programmers does it really matter that the syntax should be acceptable? I would love to know your opinion on that. Disclaimer: I'm not trying to start an argument. I thought this is a good topic of discussion. Update: This turns out to be a good topic. I'm glad you are all participating in it.

    Read the article

  • Graphical Programming Language

    - by prosseek
    In control engineering or instrumentation, I see Simulink or LabVIEW(G) is pretty popular. In ESL design, I see that Agilent SystemVue is gaining some popularity. If you see the well established compiler theroy, almost 100% is about the textual language. But how about the graphical language? Is there any noticable research or discussion about the graphical programming language? In terms of Theory about Graphical Language - syntactic/semantic analysis and whatever relevant expressiveness (Actually, I asked a question about it at SO - http://stackoverflow.com/questions/2427496/what-do-you-mean-by-the-expressiveness-in-programming-lanuguage) Possibility of the Graphical language ... Or what do you think about the Graphical Programming Language?

    Read the article

  • Recommended language and IDE for simple linux application [on hold]

    - by niklon
    I want to write a simple program on Debian with Gnome. Application will act as a side bar, giving simple information on online servers statuses. I preferably have a black transparent background(Terminal-like). I'm asking this question because I was previously writing programs in .NET C# for myself, and now I don't want to get to Mono, but something more conventional. What language should I choose for this task? What would be the recommended way to do it?(eg. what IDE)

    Read the article

< Previous Page | 48 49 50 51 52 53 54 55 56 57 58 59  | Next Page >