Search Results

Search found 3880 results on 156 pages for 'duplicate'.

Page 42/156 | < Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >

  • Who are the thought leaders in software engineering/development? [closed]

    - by Mohsin Hijazee
    Possible Duplicate: What are the big contemporary names in the programming field? I am sorry if it is a duplicate questions or is useless. I want to compile a list of influential people in our industry who can be termed as "opinionated" and thought leaders. There are basically two characteristics that I'm referring to here: The person has introduced new concepts/terminology/trends or talked about existing ones in thought provoking way. Majority or part of the writings are available online. Some of the people who I think as thought leaders are as under: Martin Fowler Known for domain specific languages, Active Record, IoC. Joel Spolsky known for his 12 point Joel test, Law of Leaky abstractions. Kent Beck known for XP. Paul Graham. Any other names and links?

    Read the article

  • Error when trying to install Menulibre in Lubuntu 12.10 [closed]

    - by cipricus
    Possible Duplicate: How can I fix a 404 Error using a PPA? When trying to add a PPA (to install menulibre) I get these errors: Err http://ppa.launchpad.net quantal/main Sources 404 Not Found Err http://ppa.launchpad.net quantal/main i386 Packages 404 Not Found W: Failed to fetch http://ppa.launchpad.net/menulibre-dev/devel/ubuntu/dists/quantal/main/source/Sources 404 Not Found W: Failed to fetch http://ppa.launchpad.net/menulibre-dev/devel/ubuntu/dists/quantal/main/binary-i386/Packages 404 Not Found Why is that and how to solve it? As the question was flagged as duplicate: I should clarify that I want to know why i cannot add a source that seems active in launchpad, and how could i add it and install the program, not just how to remove the error by removing the source in the software sources list Menulibre seems available in software sources. Why can't I install it? I'm in Lubuntu 12.10

    Read the article

  • Lazy Processing of Streams

    - by Giorgio
    I have the following problem scenario: I have a text file and I have to read it and split it into lines. Some lines might need to be dropped (according to criteria that are not fixed). The lines that are not dropped must be parsed into some predefined records. Records that are not valid must be dropped. Duplicate records may exist and, in such a case, they are consecutive. If duplicate / multiple records exist, only one item should be kept. The remaining records should be grouped according to the value contained in one field; all records belonging to the same group appear one after another (e.g. AAAABBBBCCDEEEFF and so on). The records of each group should be numbered (1, 2, 3, 4, ...). For each group the numbering starts from 1. The records must then be saved somewhere / consumed in the same order as they were produced. I have to implement this in Java or C++. My first idea was to define functions / methods like: One method to get all the lines from the file. One method to filter out the unwanted lines. One method to parse the filtered lines into valid records. One method to remove duplicate records. One method to group records and number them. The problem is that the data I am going to read can be too big and might not fit into main memory: so I cannot just construct all these lists and apply my functions one after the other. On the other hand, I think I do not need to fit all the data in main memory at once because once a record has been consumed all its underlying data (basically the lines of text between the previous record and the current record, and the record itself) can be disposed of. With the little knowledge I have of Haskell I have immediately thought about some kind of lazy evaluation, in which instead of applying functions to lists that have been completely computed, I have different streams of data that are built on top of each other and, at each moment, only the needed portion of each stream is materialized in main memory. But I have to implement this in Java or C++. So my question is which design pattern or other technique can allow me to implement this lazy processing of streams in one of these languages.

    Read the article

  • Partition showing up twice under devices

    - by JohnReed
    I have an ntfs data partition on the same drive as my ubuntu 12.04 partition. I wanted to have this partition automatically mount at startup so that dropbox can start syncing at startup without me having to manually mount it (my dropbox folder is on the partition). I edited the fstab file by appending the following to it: #Data-Partition UUID=762639DE416D0A21 /media/Data ntfs-3g rw,auto,users,exec,nls=utf8,umask=003,gid=46,uid=1000 0 2 The drive now automounts at startup like I wanted, but now I have a duplicate Data entry under my Devices list. This only appeared after editing the fstab file. The second Data entry under Devices does not mount properly and I do not want it there. Why is this duplicate entry there and how can I get rid of it? additional information: Here is the output I get from cat /proc/partitions: major minor #blocks name 8 0 976762584 sda 8 1 102400 sda1 8 2 111574016 sda2 8 3 803643392 sda3 8 4 1 sda4 8 5 55410688 sda5 8 6 6028288 sda6 11 0 1048575 sr0

    Read the article

  • Is there an eCommerce platform made to fit between a global header/footer? [closed]

    - by beta208
    Possible Duplicate: Which Ecommerce Script Should I Use? We've been looking around for a while for an eCommerce platform made to live between the header and footer for integrating in an existing site. We would prefer it not to be paypal buttons, but an actual CMS type platform. Any suggestions? This is not a duplicate, and is a valid question sought out by many across the web. If someone has an answer many people would benefit from it. This is not simply looking for a CMS with X or Y.

    Read the article

  • Trying to retrace our SEO domain redirect strategy

    - by dans
    An SEO built a copy of my company's e-commerce site on another domain that contained our product's keywords in the name (i.e. as if Levi's built a duplicate site on bluejeans.com)...and then they referenced a lot of the images on the actual website from the other domain (as if Levis.com had images on it referenced like: img src="http://www.bluejeans.com/jeans-front.jpg"), but when you tried to reach the site by typing the name into the browser you would be redirected to the regular website, so the site wasn't really used for any purpose except I guess SEO. Since I didn't think this was doing anything GOOD for us at the time, I deleted the duplicate site and let the hosting on it expire, only to watch our search engine position rankings fall dramatically. Any ideas as to what was going on there? I want to get it back to understand its impact, but I don't know how it was set up. I contacted our host and they have no idea how it was set up. I suspect there was some sort of redirect in play, or something?

    Read the article

  • Efficient way to find unique elements in a vector compared against multiple vectors

    - by SyncMaster
    I am trying find the number of unique elements in a vector compared against multiple vectors using C++. Suppose I have, v1: 5, 8, 13, 16, 20 v2: 2, 4, 6, 8 v3: 20 v4: 1, 2, 3, 4, 5, 6, 7 v5: 1, 3, 5, 7, 11, 13, 15 The number of unique elements in v1 is 1 (i.e. number 16). I tried two approaches. Added vectors v2,v3,v4 and v5 into a vector of vector. For each element in v1, checked if the element is present in any of the other vectors. Combined all the vectors v2,v3,v4 and v5 using merge sort into a single vector and compared it against v1 to find the unique elements. Note: sample_vector = v1 and all_vectors_merged contains v2,v3,v4,v5 //Method 1 unsigned int compute_unique_elements_1(vector<unsigned int> sample_vector,vector<vector<unsigned int> > all_vectors_merged) { unsigned int duplicate = 0; for (unsigned int i = 0; i < sample_vector.size(); i++) { for (unsigned int j = 0; j < all_vectors_merged.size(); j++) { if (std::find(all_vectors_merged.at(j).begin(), all_vectors_merged.at(j).end(), sample_vector.at(i)) != all_vectors_merged.at(j).end()) { duplicate++; } } } return sample_vector.size()-duplicate; } // Method 2 unsigned int compute_unique_elements_2(vector<unsigned int> sample_vector, vector<unsigned int> all_vectors_merged) { unsigned int unique = 0; unsigned int i = 0, j = 0; while (i < sample_vector.size() && j < all_vectors_merged.size()) { if (sample_vector.at(i) > all_vectors_merged.at(j)) { j++; } else if (sample_vector.at(i) < all_vectors_merged.at(j)) { i++; unique ++; } else { i++; j++; } } if (i < sample_vector.size()) { unique += sample_vector.size() - i; } return unique; } Of these two techniques, I see that Method 2 gives faster results. 1) Method 1: Is there a more efficient way to find the elements than running std::find on all the vectors for all the elements in v1. 2) Method 2: Extra overhead in comparing vectors v2,v3,v4,v5 and sorting them. How can I do this in a better way?

    Read the article

  • MySQL Master - Master Broken

    - by Recc
    I've Inherited a Mysql master master system, I've noticed the second master (lets call it slave from now on as it's running on a 'slave' machine) stopped getting its db's updated. I saw that Master: Slave_IO_Running: Yes Slave_SQL_Running: Yes Slave: (with an error I truncated) Slave_IO_Running: Yes Slave_SQL_Running: No Last_Errno: 1062 Last_Error: Error 'Duplicate entry '3' for key 'PRIMARY'' on [...] I don't know what caused it to process considering we cant get duplicate there. What's important is to resume normal operations; Right now I've stop slave; on the Master and stop slave; on the Slave because I saw that if I change records on the Slave the changes Do Get Propagated to Master which is in active use. How do I: Force sync EVERYTHING from master to slave without affecting data on master? Then hopefully have slave pickup replication as usual? UPDATE OK I Tried deleting all tables on slave then it complained in that error section that the 'table' doesnt exist. So i made a no data dump of Master, and made sure I have only empty tables in Secondary (slave). I start slave; on slave BUT now it's complaining about bloody alter table statements for instance: Last_Errno: 1060 Last_Error: Error 'Duplicate column name [...] Query: 'ALTER TABLE [...] How to skip the fracking alter statements I just want to replicate the bloody data and be done with it, my tables have the lates changes already FFS and now its complaining about changes made after the replication seized weeks ago How do I reset the log or something? OUTSTANDING Why would this start happening? The "Secondary" is propagating to "Primary". "Primary" is not propagating to "Secondary". But any fixes I tried to do left it in the same state Yes-Yes Yes-No with same Last_Error. I think around that time the server was taken off the network, could that confuse MySQL in some way?

    Read the article

  • Is this normal or does my AVG anitivirus have a virus? [closed]

    - by user390480
    Possible Duplicate: Computer is infected by a virus or a malware, what do I do now? [Note: This is not a duplicate and should not be closed as a duplicate as this is nothing like the other question. I am not asking if I have a virus, nor am I asking "what do I do now". I know that I have a virus and I know what to do. However, I am asking if these are normal AVG ads or if it has been taken over.] My Windows XP PC started acting strange and I am right now actually in Linux running off of a USB drive. I am running Avast under Linux and it has discovered some viruses on my XP drive. Some of the strange things happening in XP were: I could not get to Google.com My Hosts file was set to hidden and read only My Hosts file had an entry of ::1 And AVG had ads in it I've never seen before. Maybe it is normal but I Binged for AVG anti-virus and become.com but found no information. (The red lines and question mark are by me) Any thoughts?

    Read the article

  • Retrieve Deleted Project in MS Visual SourceSafe?

    - by AJ
    A while ago, I "tidied up" our Visual SourceSafe repository, including deleting a duplicate of one project. It now turns out that the duplicate had some code I need to see. How can I get back a deleted project? I can't see from any of the docs how to do this. Thanks for any help.

    Read the article

  • Linq to object - Select Distinct

    - by Ben
    Hi, I can't quite figure out why this Linq Statement isn't working as i would expect: Dim distinctSurchargesList = (From thisparent As Parent In ThisParentCollection _ From thisChild As Child In thisparent.theseChildren _ Select New With {.childId = thischild.Id}).Distinct() I would assume that this would create a new collection of anonymous types, that would be distinct. Instead it creates a collection the size of the "ThisParentCollection" with duplicate "MyAnonymousType" in it (duplicate id's). Can anyone tell me where im going wrong? Thanks

    Read the article

  • Javascript source code analysis ( specifically duplication checking )

    - by David
    Partial duplicate of this Notes: I already use JSLint extensively via a tool I wrote that scans in intervals my current project directory for recently updated/created .js files. It's drastically improved productivity for me and I doubt there is anything as good as JSLint for the price (it's free). That said, is there any analysis tool out there that can find repetitive or near-duplicate code blocks, the goal being to make it easier to find opportunities to consolidate large files or small/medium sized projects?

    Read the article

  • Developer name on the App Store

    - by aon
    I feel worry when i read "company name cannot be edited once submitted." Is it possible that i submit my developer name duplicate to other? (same name) I do not want my developer name duplicate to other. Sorry for my english.

    Read the article

  • WCF v.s. legacy ASP.Net Web Services

    - by George2
    Duplicate: although this is a good discussion, this is a duplicate of Web Services — WCF vs. Standard. Please consider adding any new information to the earlier question and closing this one. Could anyone recommend me some documents to describe why WCF is better than legacy ASP.Net web services? I am especially interested in performance and security. Thanks!

    Read the article

  • How can I prevent a double submit with jQuery or Javascript?

    - by kielie
    Hi guys, I keep getting duplicate entries in my database because of impatient users clicking the submit button multiple times. I googled and googled and found a few scripts, but none of them seem to be sufficient. How can I prevent these duplicate entries from occurring using javascript or preferably jQuery? Thanx in advance!

    Read the article

  • Tool to recursively search all files in a directory for a string [closed]

    - by routeNpingme
    Possible Duplicate: Tools to search for strings inside files without indexing Before I make one, any good tools out there for free/cheap that will recurse through all files in a directory and search for a text string in them? Need to find an instance of a string in any configuration files laying around. Kind of like a Windows Search "find files that contain..." on steroids? Oops - This is pretty much a duplicate question, sorry, didn't find the other one before

    Read the article

  • Problem in java.util.Set.addAll() method

    - by Yatendra Goel
    I have a java.util.Set<City> cities and I need to add cities to this set in 2 ways: By adding individual city (with the help of cities.add(city) method call) By adding another set of cities to this set (with the help of cities.addAll(anotherCitiesSet) method call) But the problem in second approach is that i don't know whether there were any duplicate cities in the anotherCitiesSet. I want to do some processing whenever a duplicate entry is tried to be entered in thecities set.

    Read the article

  • unable to install latest phpUnit in Ubuntu 10.04

    - by Dex Barrett
    I'm trying to install PHPUnit in Ubuntu 10.04 but I get these error messages sudo pear install -a pear.phpunit.de/PHPUnit Duplicate package channel://pear.phpunit.de/File_Iterator-1.3.3 found Duplicate package channel://pear.phpunit.de/File_Iterator-1.3.2 found install failed I tried reinstalling PEAR, upgrading it; updated the PEAR and PHPUnit channel; cleared the PEAR's cache but still no luck, I keep getting the same error. Does anyone have the same problem and know how to solve it? Thank you.

    Read the article

  • Generating report with MySQL and Rails - how?

    - by Arywista
    Here is my data model from my application: id :integer(4) not null, primary key spam :boolean(1) not null duplicate :boolean(1) not null ignore :boolean(1) not null brand_id :integer(4) not null attitude :string not null posted_at :datetime not null Attitude could have 3 states: negative, positive, neutral. I want to generate resultset in table, this way, for each day between start and end date: date | total | positive | neutral | negative 2009-10-10 | 12 | 4 | 7 | 1 (...) 2009-10-30 | 5 | 2 | 1 | 1 And ignore all records which have: duplicate = true ignore = true spam = true How it's could be done?

    Read the article

< Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >