Search Results

Search found 3618 results on 145 pages for 'huge'.

Page 44/145 | < Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >

  • Resizing a rectangular array

    - by devdept
    Does a smarter way than the following exist to resize a rectangular array? double[,] temp = new double[newSize, originalSecondDimension]; Array.Copy(original, temp, original.Length); I was concerned about duplicating a huge array and the memory necessary to do it. What does the Array.Resize() do internally? Thanks, Alberto

    Read the article

  • can i quickly run entire page's text through a function on page load?

    - by korben
    i have setup a profanity filter with bad words in a XML file and have the following function to run on my page to replace the words: BadWordFilter.Instance.GetCleanString(TextBox1.Text); i'm about to go through my entire site now wrapping that function around every little text variable one by one and it's going to be a huge pain in the butt i'm hoping there's a way that i could just set my masterpage to automatically run all text through this thing on any page_load, so that the effect would be site-wide instantly. is this possible? much appreciated for any help

    Read the article

  • Scrum backlog sizing is taking forever

    - by zachary
    I work on a huge project. While we program we end up meeting for endless backlog sizing sessions where all the developers sit down with the team and size user stories. Scrum doubters are saying that this process is taking too long and development time is being wasted. My question is how long should it take to size a user story on average? And does anyone have any tips to make these sizing sessions go quicker?

    Read the article

  • python cairoplot store previous readings..

    - by krisdigitx
    hi, i am using cairoplot, to make graphs, however the file from where i am reading the data is growing huge and its taking a long time to process the graph is there any real-time way to produce cairo graph, or at least store the previous readings..like rrd. -krisdigitx

    Read the article

  • Database maintenance, big binary table optimization

    - by jgemedina
    Hi i have a huge database, around 1 TB in size, most of the space is consumed by a table which stores images, the tables has right now almost 800k rows. server response time has increased, i would like to know which techniques should i use or you recomend, partitioning? o how to reorganize the table every row is accessed by the image id column, and it has its clustered index by that column, and every two days i reorganize the index and every 7 days i rebuild it, but it seems not to be working any suggestions?

    Read the article

  • Random record from MongoDB

    - by Will M
    I am looking to get a random record from a huge (100 million record) mongodb. What is the fastest and most efficient way to do so? The data is already there and there are no field in which I can generate a random number and obtain a random row. Any suggestions?

    Read the article

  • Visual Studio Generating Code in Release mode

    - by bball
    I am using Visual Studio to compile some C++ Code. When I compile my program in Debug mode it works perfectly, but when I try compiling it in Release mode, VS halts when Generating Code (ie: took 20 mins and still didn't finish). After I forced VS to stop the build process, the executable generated was huge 2MB while it should be about 304KB. Please can you help figure this out. How should I setup the configuration parameters for Release mode to work? Thanks,

    Read the article

  • Advantage Database Server: in-memory queries.

    - by ie
    As far as I know, ADS v.10 tries to keep result of query in memory until it is a quite huge. The same should be true for the __output table and for temporary tables. When the result becoming large, swapping stated. The question is what memory limit is set for a query, a worker, whatever? Could this limit be configured? Thanks.

    Read the article

  • Compiling VIM on OS X Mountain Lion

    - by Lauri L
    Compiling VIM on Mountain Lion does not work for me. I compile VIM with the following sequence: hg clone https://vim.googlecode.com/hg/ vim cd vim ./configure --enable-rubyinterp --with-features=huge -prefix=/usr make I then try to start the compiled binary with src/vim, but it crashes like this: $ src/vim Vim: Caught deadly signal SEGV Vim: Finished. Segmentation fault: 11 What am I missing here?

    Read the article

  • PHP Filesystem Pagination

    - by Byron
    How does one paginate a large listing of files within a folder? I can't see any functions in the PHP documentation that mention any way to specify an 'offset'. Both glob() and scandir() simply return all the files in the folder, and I'm afraid that won't be a good idea for a huge directory. Is there any better way of doing this than simply going through all the files and chopping off the first X number of files?

    Read the article

  • Complex query with two tables and multilpe data and price ranges

    - by TiuTalk
    Let's suppose that I have these tables: [ properties ] id (INT, PK) name (VARCHAR) [ properties_prices ] id (INT, PK) property_id (INT, FK) date_begin (DATE) date_end (DATE) price_per_day (DECIMAL) price_per_week (DECIMAL) price_per_month (DECIMAL) And my visitor runs a search like: List the first 10 (pagination) properties where the price per day (price_per_day field) is between 10 and 100 on the period for 1st may until 31 december I know thats a huge query, and I need to paginate the results, so I must do all the calculation and login in only one query... that's why i'm here! :)

    Read the article

  • Subdomain forwarding using .htaccess

    - by RJ
    I want to redirect a praticular subdomain to the main domain http(s)://dl.example.com/par1/par2 to http(s)://www.example.com/par1/par2 How to achieve the above using .htaccess Why i want to do this: Whenever any user download a file from my server, if the file is huge , then user cannot do any other operation until the file is downloaded completely...so the solution that i have thought is to forward the download request through subdomain so that the browser may continue with rest of the operation. Thanks

    Read the article

  • Formatting equations in LaTeX

    - by jetsam
    When I include an equation in LaTeX that is enumerated, i.e. {\begin{equation} $$ $$ ... \end{equation} } The line above the equation (blank space between the text preceeding it and the equation) is huge. How do I make it smaller?

    Read the article

  • SVN Merge Question

    - by SVN
    Suppose that I have two massive folders in SVN: test and prod. I'm trying to use the following svn command to merge the content from test into prod: svn merge ./prod@HEAD ./test@HEAD ./prod However, my SVN repository is huge and this process takes an enormous amount of time. Is there any way that I can make a simpler merge statement which will be more processor friendly? Perhaps a server-URL based merge? Any insight is appreciated....

    Read the article

  • how to save a fitted R model for later use

    - by ahala
    Sorry for this novice question: if I fit a lm() model or loess() model, and save the model somewhere in a file or in database, for later using by third party with predict() method, do I have to save the entire model object? Since returned model object contains orginal raw data, this returned object can be huge.

    Read the article

  • crowd website simulation on localhost for a php/mysql project

    - by Mac Taylor
    hey guys I searched for a while on how to find a benchmarking software that can simulate crowd website with more than 1000 users online to find out leaks in my php/mysql script . as long as i ran my script for a huge community and it wasn't successful enough and lots of RAM usage happened , now I need a way to simulate that much usage to benchmark my script and optimize it . I am using XAMMP Local Server and my project written in PHP&MYSQL. thanks in advance

    Read the article

  • C - How to implement Set data structure?

    - by psihodelia
    Is there any tricky way to implement a set data structure (a collection of unique values) in C? All elements in a set will be of the same type and there is a huge RAM memory. As I know, for integers it can be done really fast'N'easy using value-indexed arrays. But I'd like to have a very general Set data type. And it would be nice if a set could include itself.

    Read the article

  • Count the number of lines in a file with Ruby, without reading entire file into memory

    - by smnirven
    I'm processing huge data files (millions of lines each). Before I start processing I'd like to get a count of the number of lines in the file, so I can then indicate how far along the processing is. I am using Ruby, and because of the size of the files, it would not be practical to read the entire file into memory just to count how many lines there are. Does anyone have a good suggestion on how to do this?

    Read the article

  • C: performance of assignments, binary operations, et cetera...

    - by Shinka
    I've heard many things about performance in C; casting is slow compared to normal assignments, functional call is slow, binary operation are much faster than normal operations, et cetera... I'm sure some of those things are specific to the architecture, and compiler optimization might make a huge difference, but I would like to see a chart to get a general idea what I should do and what I should avoid to write high-performance programs. Is there such a chart (or a website, a book, anything) ?

    Read the article

< Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >