Search Results

Search found 17602 results on 705 pages for 'computer engineering'.

Page 3/705 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • The Best Ways to Make Use of an Idle Computer

    - by Lori Kaufman
    If you leave your computer on when you are not using it, there are ways you can put your computer to use when it’s sitting idle. It can do scientific research, backup your data, and even look for signs of extraterrestrial life. How To Create a Customized Windows 7 Installation Disc With Integrated Updates How to Get Pro Features in Windows Home Versions with Third Party Tools HTG Explains: Is ReadyBoost Worth Using?

    Read the article

  • If Computer Problems were Physical Life Events [Video]

    - by Asian Angel
    Things can be bad (and frustrating) enough when you have problems with your computer, but what if those events actually crossed over into physical reality? Note: Video contains some language that may be considered inappropriate. If Computer Problems Were Real – Awkward Spaceship [via Fail Desk] How To Properly Scan a Photograph (And Get An Even Better Image) The HTG Guide to Hiding Your Data in a TrueCrypt Hidden Volume Make Your Own Windows 8 Start Button with Zero Memory Usage

    Read the article

  • How is Basic Physics applied in CS/SE?

    - by Wulf
    What basic physics principles do software engineers and/or computer scientists use to help solve specific or common problems? The first one that came to my head was creating a Physics engine for a game; physics is involved, as it requires knowledge of: Forces and Motion: Kinematics, Dynamics, Circular Motion However, I need another example, but haven't come across one that involves basic physics. Please consider the following basic physics (grade 12 level) concepts: Energy and Momentum: Work and Energy, Momentum and Collisions, Gravitational and Celestial Mechanics Electric, Gravitational & Magnetic Field: Electric Charges and Electric Field, Magnetic Fields and Electomagnetism The Wave Nature of Light: Waves and Light, Wave Effects of Light Matter-Energy Interface: Einstein’s Special Theory of Relativity, Waves, Photons and Matter, Radioactivity and Elementary Particles I will be happy with any response; Keywords for google, names of methods like raycasting, etc.

    Read the article

  • Is the science of Computer Science dead?

    - by Veaviticus
    Question : Is the science and art of CS dead? By that I mean, the real requirements to think, plan and efficiently solve problems seems to be falling away from CS these days. The field seems to be lowering the entry-barrier so more people can 'program' without having to learn how to truly program. Background : I'm a recent graduate with a BS in Computer Science. I'm working a starting position at a decent sized company in the IT department. I mostly do .NET and other Microsoft technologies at my job, but before this I've done Java stuff through internships and the like. I personally am a C++ programmer for my own for-fun projects. In Depth : Through the work I've been doing, it seems to me that the intense disciplines of a real science don't exist in CS anymore. In the past, programmers had to solve problems efficiently in order for systems to be robust and quick. But now, with the prevailing technologies like .NET, Java and scripting languages, it seems like efficiency and robustness have been traded for ease of development. Most of the colleagues that I work with don't even have degrees in Computer Science. Most graduated with Electrical Engineering degrees, a few with Software Engineering, even some who came from tech schools without a 4 year program. Yet they get by just fine without having the technical background of CS, without having studied theories and algorithms, without having any regard for making an elegant solution (they just go for the easiest, cheapest solution). The company pushes us to use Microsoft technologies, which take all the real thought out of the matter and replace it with libraries and tools that can auto-build your project for you half the time. I'm not trying to hate on the languages, I understand that they serve a purpose and do it well, but when your employees don't know how a hash-table works, and use the wrong sorting methods, or run SQL commands that are horribly inefficient (but get the job done in an acceptable time), it feels like more effort is being put into developing technologies that coddle new 'programmers' rather than actually teaching people how to do things right. I am interested in making efficient and, in my opinion, beautiful programs. If there is a better way to do it, I'd rather go back and refactor it than let it slide. But in the corporate world, they push me to complete tasks quickly rather than elegantly. And that really bugs me. Is this what I'm going to be looking forward to the rest of my life? Are there still positions out there for people who love the science and art of CS rather than just the paycheck? And on the same note, here's a good read if you haven't seen it before The Perils Of Java Schools

    Read the article

  • Computer Science Degrees and Real-World Experience

    - by Steven Elliott Jr
    Recently, at a family reunion-type event I was asked by a high school student how important it is to get a computer science degree in order to get a job as a programmer in lieu of actual programming experience. The kid has been working with Python and the Blender project as he's into making games and the like; it sounds like he has some decent programming chops. Now, as someone that has gone through a computer science degree my initial response to this question is to say, "You absolutely MUST get a computer science degree in order to get a job as a programmer!" However, as I thought about this I was unsure as to whether my initial reaction was due in part to my own suffering as a CS student or because I feel that this is actually the case. Now, for me, I can say that I rarely use anything that I learned in college, in terms of the extremely hard math, algorithms, etc, etc. but I did come away with a decent attitude and the willingness to work through tough problems. I just don't know what to tell this kid; I feel like I should tell him to do the CS degree but I have hired so many programmers that majored in things like English, Philosophy, and other liberal arts-type degrees, even some that never went to college. In fact my best developer, falls into this latter category. He got started writing software for his church or something and then it took off into a passion. So, while I know this is one of those juicy potential down vote questions, I am just curious as to what everyone else thinks about this topic. Would you tell a high school kid about this? Perhaps if he/she already knows a good deal of programming and loves it he doesn't need a CS degree and could expand his horizons with a liberal arts degree. I know one of the creators of the Django web framework was a American Literature major and he is obviously a pretty gifted developer. Anyway, thanks for the consideration.

    Read the article

  • Proof Identify stolen computer getting computer identification info from Launchpad bugs and comparing

    - by Kangarooo
    I sold my old laptop to neighbours and it was stolen from them. Well i think i have found thief so i want to check his computer id and compare it to my old Launchpad bugs id. How in Launchpad i can find from my bugs: Motherboard HDD Somthing else that can help identify it Maybe how to recover or find some overwritten files (couse now there is windows) I found in Launchpad one my bugs has LSPCI autogenerated from bug 682846 https://launchpadlibrarian.net/70611231/Lspci.txt but i dont see any id that can be used to identify specificly my comp. This can be used to identify many same models. Or i missed something in there? And what commands should i use to get all identification on that comp in one go fast? Just lspci? How to get same lspci as it is in that Launchpad link? Now testing laspci on my computer i dont get so much info. Also im now doing a search in my external hdd where i have many backups and maybe i have there result from lspci. So what containing keywords would help doing search with for small lspci and full reports ive done? I might have done sudo lshw somefilename

    Read the article

  • Turing Machine & Modern Computer

    - by smwikipedia
    I heard a lot that modern computers are based on Turing machine. I'd like to share my understanding and hear your comments. I think the computer is a big general-purpose Turing machine. Each program we write is a small specific-purpose Turing machine. The classical Turing machine do its job based on the input and its current state inside and so do our programs. Let's take a running program (a process) as an example. We know that in the process's address space, there's areas for stack, heap, and code. A classical Turing machine doesn't have the ability to remember many things, so we borrow the concept of stack from the push-down automaton. The heap and stack areas contains the state of our specific-purpose Turing machine (our program). The code area represents the logic of this small Turing machine. And various I/O devices supply input to this Turing machine. The above is my naive understanding about the working paradigm of modern computer. I couln't wait to hear your comments. Thanks very much.

    Read the article

  • Do the ideas of traditional software engineering conflict with the newer agile development technique

    - by fuentesjr
    So as a developer I am seeking to improve not only my coding skills but my design and management skills. Because of this I'm starting to pay more attention to software engineering practices but i'm not sure where agile development fits into the picture. I can appreciate agility in projects but I wonder whether this conflicts with the traditional ways of software engineering practices and research.

    Read the article

  • How will my Electronic Engineering degree be received in the Canadian Game Development market? [closed]

    - by Harikawashi
    I have a Electronic Engineering with Computer Science Degree from a reputable South African university. The EE with CS degree is basically Electronic Engineering, with some of the high voltage subjects thrown out and replaced with computer science subjects - mostly quite theoretical, but not in too much depth. I went on to earn a Masters Degree in Digital Signal Processing, focussing on Speech Recognition in Educational Applications. I have always loved programming - I taught myself QBASIC when I was in primary school, I learned Java at school, did some low level C at University, and taught myself C# and Python while doing my post graduate degree. C# is currently my strong suit, I think I am pretty capable with it. I have two years work experience in Namibia - working as a consulting electrical engineer (no software content whatsoever) and also developing C# desktop applications for the company I work for. I would like to move to Canada next year and work in the Game Development Industry as programmer or software engineer. My interests in particular are towards the more mathematical applications, like game and physics engines, or statistical disciplines like artificial intelligence. However, these are passions - not areas in which I have any work experience. So the question: How well will my BEngEE&CS and MScEng be received in the game industry? Seeing as it's not a pure software degree and I have no official software development work experience?

    Read the article

  • Will people respect a Masters of Science in IT w/software engineering concentration from RPI?

    - by twneale
    Here's my thing: I got my undergraduate degree in political science, then a law degree. Then I figured out that I love programming and I'm pretty good at it too. It's fun and rewarding enough for me that I'd prefer to do it for a living over almost any form of pure law practice. So I'm looking at getting a masters degree to put some weight behind a possible career switch. If I actually want to develop software (web, in particular), would people in programming circles respect a master's of science in IT? Specifically, consider as an example the MS in IT from Rensselaer Polytechnic Institute (with a concentration in software engineering). Here's the home page: http://www.rpi.edu/IT/graduate/masters_program.html In particular, I mean to draw a contrast between IT as specifically contemplated by the RPI masters program (an interdisciplinary tech/business program) and other MS degrees in computer science or software engineering that focus more on the science and technical aspects. I guess I want to make sure that other programmers would respect my credentials and not consider me as different or underqualified based on the connotations of the phrase "IT". I believe RPI has an unimpeachable reputation for hard science, and the program seems excellent, but it still matters to me how people in industry would perceive it.

    Read the article

  • Learning computer architecture as a programmer

    - by Samaursa
    I typically run across gurus at SO and other places (instructors, book authors etc.) that would say something along the lines "This will cause alignment issues" or other low level tidbits. I want to learn about all these tidbits that are relevant to programming. Now usually when I see low level books (computer architecture books for example) they are too low level and geared towards people whose primary area of interest is computer architecture and not software design. Do you have recommendations for books that go through low-level stuff that is relevant to programmers?

    Read the article

  • Computer vision algorithms (how is this possible?)

    - by Maxim Gershkovich
    I recently stumbled across a company that has created what appears to be a computer vision technology that is capable of detecting shoplifting automatically and alert its users. LINK Watching some of the videos and examples provided by the company has left me completely baffled and amazed as to how on earth they may have achieved this functionality. I understand that no-one here will be able to tell me exactly how this may have been achieved but is anyone aware - and could point me to - research in this field or alternatively perhaps provide details as to how something like this could be implemented or guidance of where one might start? My understanding was the computer vision algorithms were many years away from being this sophisticated. Is this sort of application really possible? Anyone willing to hazard a guess at how they achieved this?

    Read the article

  • Master Degree in MIS for computer science student

    - by tnhan07
    I'm junior student in computer science. After taking half of my major related courses, I found that I don't like this theoretical side of IT. As a result, I decided that I would devote my career to CIS/MIS because it is more interesting. However, some veteran programmers in this forum said that having a strong computer science foundation would help much for CIS. Therefore, I think it's better for me to complete my CS degree then have a Master Degree in MIS than have a minor in MIS. After some internet searching, I found that top universities(in my reach) offering master degree in CIS/MIS are all business schools, is there any obstacle for a CS student who lacks of business knowledge like me if I study in these schools? Do you have any advice for me?

    Read the article

  • Good Intro to Computer Science Book for FE Developer [on hold]

    - by Squirkle
    I am a JavaScript developer/architect who, like many developers these days, did not come from a Computer Science background (I studied Philosophy at a liberal arts college), but instead learned development by actually building applications, and by reading books explaining language grammars, design patterns, and best practices. I have never felt that my ignorance of CS concepts has hurt my ability to build great apps or find employment. Recently, however, I have felt the itch to grow in this direction. Do you have any suggestions for some good introductory CS resources/books? I know that Computer Science is a huge field and my question is very general, but I am looking for a 101-type survey of the high-level concepts, from which I can branch off into more specific areas of study. Thanks!

    Read the article

  • Master's in Software Engineering vs. Master's in Computer Science: which degree is preferred by empl

    - by dbarker
    I've been building software professionally for 7 years and am considering a master's degree. I understand the difference between these two degrees as simply: MSCS is the theory while MSE is the practice. I'm equally interested in both and would be happy with either, although I'm curious how these degrees rank in the eyes of a potential employer. I could see two views that a hiring manager could possibly take: a MSCS is loftier and has an implied knowledge of Software Engineering an MSE is more practical and has an implied knowledge of Computer Science In my own experience I've seen both MSCS degree holders than cannot program at all while others are among the best programmers I've met, so of course actual ability will depend on the individual. My question is about the "on paper" value of these two degrees when seeking a job. All things considered, is one degree more hirable, higher-paying than the other?

    Read the article

  • Super Computer Built from Raspberry Pi Boards and LEGO Bricks

    - by Jason Fitzpatrick
    It was only a matter of time before someone chained together dozens of Raspberry Pi boards into a serviceable super computer; read on to see how a team of Southampton scientists built a 64-core machine using them. Image courtesy of Simon Cox and the University of Southampton. From the University of South Hampton press release: Professor Cox comments: “As soon as we were able to source sufficient Raspberry Pi computers we wanted to see if it was possible to link them together into a supercomputer. We installed and built all of the necessary software on the Pi starting from a standard Debian Wheezy system image and we have published a guide so you can build your own supercomputer.” The racking was built using Lego with a design developed by Simon and James, who has also been testing the Raspberry Pi by programming it using free computer programming software Python and Scratch over the summer. The machine, named “Iridis-Pi” after the University’s Iridis supercomputer, runs off a single 13 Amp mains socket and uses MPI (Message Passing Interface) to communicate between nodes using Ethernet. The whole system cost under £2,500 (excluding switches) and has a total of 64 processors and 1Tb of memory (16Gb SD cards for each Raspberry Pi). Professor Cox uses the free plug-in ‘Python Tools for Visual Studio’ to develop code for the Raspberry Pi. How to Get Pro Features in Windows Home Versions with Third Party Tools HTG Explains: Is ReadyBoost Worth Using? HTG Explains: What The Windows Event Viewer Is and How You Can Use It

    Read the article

  • What are some good Software Engineering books for people who didn't formally study Computer Science or Software Engineering?

    - by Kugathasan Abimaran
    I'm a graduate in the electronic & telecommunication field, but working in a software company. I want to continue in this field and going for Masters in it. Can you recommend me some of the best books on software engineering, which cover almost all the topics in software engineering. I am not looking for books about coding practices such as Code Complete, Pragmatic Programmer, but rather general software engineering references.

    Read the article

  • Which is the next dominant programming paradigm? [closed]

    - by Kugathasan Abimaran
    What is the next programming paradigm when OOP get lost in the market? Or else will OOP be for ever? What is your advise for the future developers? To which paradigm should we aware of? Because, before OOP, structured programming paradigm is there with C. Don't close it Please, because I need to aware, which paradigm have the ability to withstand in future? Aspect-oriented programming. Declarative programming. Functional programming. Object-oriented programming. Any Others? This describes programming paradigm according to their kernel language.

    Read the article

  • IBM sort Rational Engineering Lifecycle Manager et SmartCloud Continuous Delivery pour réduire la complexité des applications

    IBM sort Rational Engineering Lifecycle Manager et SmartCloud Continuous Delivery des nouveaux outils pou réduire la complexité des applications Le portefeuille d'outils de développement d'IBM vient de s'enrichir de Rational Engineering Lifecycle Manager et SmartCloud Continuous Delivery, deux nouvelles solutions pour répondre aux besoins des développeurs. Ces outils ont été conçus pour résoudre des problèmes rencontrés dans les projets d'envergure comme ceux qu'on trouve dans des systèmes pour l'industrie automobile, la robotique ou de petits changements dans le système qui peuvent avoir des effets plus importants que prévu initialement. La solution Rational Engineering L...

    Read the article

  • How to Tell If Your Computer is Overheating and What to Do About It

    - by Chris Hoffman
    Heat is a computer’s enemy. Computers are designed with heat dispersion and ventilation in mind so they don’t overheat. If too much heat builds up, your computer may become unstable or suddenly shut down. The CPU and graphics card produce much more heat when running demanding applications. If there’s a problem with your computer’s cooling system, an excess of heat could even physically damage its components. Is Your Computer Overheating? When using a typical computer in a typical way, you shouldn’t have to worry about overheating at all. However, if you’re encountering system instability issues like abrupt shut downs, blue screens, and freezes — especially while doing something demanding like playing PC games or encoding video — your computer may be overheating. This can happen for several reasons. Your computer’s case may be full of dust, a fan may have failed, something may be blocking your computer’s vents, or you may have a compact laptop that was never designed to run at maximum performance for hours on end. Monitoring Your Computer’s Temperature First, bear in mind that different CPUs and GPUs (graphics cards) have different optimal temperature ranges. Before getting too worried about a temperature, be sure to check your computer’s documentation — or its CPU or graphics card specifications — and ensure you know the temperature ranges your hardware can handle. You can monitor your computer’s temperatures in a variety of different ways. First, you may have a way to monitor temperature that is already built into your system. You can often view temperature values in your computer’s BIOS or UEFI settings screen. This allows you to quickly see your computer’s temperature if Windows freezes or blue screens on you — just boot the computer, enter the BIOS or UEFI screen, and check the temperatures displayed there. Note that not all BIOSes or UEFI screens will display this information, but it is very common. There are also programs that will display your computer’s temperature. Such programs just read the sensors inside your computer and show you the temperature value they report, so there are a wide variety of tools you can use for this, from the simple Speccy system information utility to an advanced tool like SpeedFan. HWMonitor also offer this feature, displaying a wide variety of sensor information. Be sure to look at your CPU and graphics card temperatures. You can also find other temperatures, such as the temperature of your hard drive, but these components will generally only overheat if it becomes extremely hot in the computer’s case. They shouldn’t generate too much heat on their own. If you think your computer may be overheating, don’t just glance as these sensors once and ignore them. Do something demanding with your computer, such as running a CPU burn-in test with Prime 95, playing a PC game, or running a graphical benchmark. Monitor the computer’s temperature while you do this, even checking a few hours later — does any component overheat after you push it hard for a while? Preventing Your Computer From Overheating If your computer is overheating, here are some things you can do about it: Dust Out Your Computer’s Case: Dust accumulates in desktop PC cases and even laptops over time, clogging fans and blocking air flow. This dust can cause ventilation problems, trapping heat and preventing your PC from cooling itself properly. Be sure to clean your computer’s case occasionally to prevent dust build-up. Unfortunately, it’s often more difficult to dust out overheating laptops. Ensure Proper Ventilation: Put the computer in a location where it can properly ventilate itself. If it’s a desktop, don’t push the case up against a wall so that the computer’s vents become blocked or leave it near a radiator or heating vent. If it’s a laptop, be careful to not block its air vents, particularly when doing something demanding. For example, putting a laptop down on a mattress, allowing it to sink in, and leaving it there can lead to overheating — especially if the laptop is doing something demanding and generating heat it can’t get rid of. Check if Fans Are Running: If you’re not sure why your computer started overheating, open its case and check that all the fans are running. It’s possible that a CPU, graphics card, or case fan failed or became unplugged, reducing air flow. Tune Up Heat Sinks: If your CPU is overheating, its heat sink may not be seated correctly or its thermal paste may be old. You may need to remove the heat sink and re-apply new thermal paste before reseating the heat sink properly. This tip applies more to tweakers, overclockers, and people who build their own PCs, especially if they may have made a mistake when originally applying the thermal paste. This is often much more difficult when it comes to laptops, which generally aren’t designed to be user-serviceable. That can lead to trouble if the laptop becomes filled with dust and needs to be cleaned out, especially if the laptop was never designed to be opened by users at all. Consult our guide to diagnosing and fixing an overheating laptop for help with cooling down a hot laptop. Overheating is a definite danger when overclocking your CPU or graphics card. Overclocking will cause your components to run hotter, and the additional heat will cause problems unless you can properly cool your components. If you’ve overclocked your hardware and it has started to overheat — well, throttle back the overclock! Image Credit: Vinni Malek on Flickr     

    Read the article

  • Windows 7 Computer Management Not Accessible

    - by David
    I am having issues trying to access Computer Management from the start menu. The following is a screenshot of exactly how I want to go to computer management. I know that there are other ways to get to computer management, but I would like to go the easier way through the start menu. I tried restarting the computer and I tried patching the computer. No error message comes up or anything. There is no response from the computer.

    Read the article

  • How to understand computer science as a whole?

    - by MrCellophane
    I am a college student in Computer Science, I have been studying CS for a long time. But even till today, I still feel so confused about a lot of things. First of all, I have solid foundation in alg, data structure, and OOP, but I don't have a clear understanding of the whole system of the subject. I studied database, alg, data structure, OOP, network, software engineering, OS, and a lot of courses. I know what they are, but I don't know how to connect them together. Is there a way to have a clear understanding of the architecture of the subject? And is there a way to know what technology is used to do what? For example, in interviews, when people ask me alg, data str, Java, OS or any other specific field, I could answer, but when they ask some other very general about the field, I have no idea. Well, I know my question maybe a little bit confusing, but what my situation is I don't even know how to ask a clear question. I don't know my question, it's totally a mess in my head. Is there a way to make it clear?

    Read the article

  • Considering getting into reverse engineering/disassembly

    - by Zombies
    Assuming a decent understanding of assembly on common CPU architectures (eg: x86), how can one explore a potential path (career, fun and profit, etc) into the field of reverse engineering? There is so little educational guides out there so it is difficult to understand what potential uses this has today (eg: is searching for buffer overflow exploits still common, or do stack monitoring programs make this obselete?). I am not looking for any step by step program, just some relevant information such as tips on how to efficiently find a specific area of a program. Basic things in the trade. As well as what it is currently being used for today. So to recap, what current uses does reverse engineering yield today? And how can one find some basic information on how to learn the trade (again it doesn't have to be step-by-step, just anything which can through a clue would be helpful).

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >