Search Results

Search found 29638 results on 1186 pages for 'phone number'.

Page 329/1186 | < Previous Page | 325 326 327 328 329 330 331 332 333 334 335 336  | Next Page >

  • Find keyword values from PDF [closed]

    - by JukkaA
    I have a lot of PDF reports I'd need to index. They're mostly "text-based PDFs", not images. I know they all have account number in certain format, 123456AAAAA and some other keyword info like addresses, customer names etc. needed in indexing these files. Basically if the file is ab.pdf, I need to create ab.txt that contains: ACC=123456AAAA Customer=John Doe Date=20120808 What would be the best software/solution to generate indexing information for these? I know there's pdftotext, but piping it to different grep/awk commands is a hack... It would be nice to specify an area in PDF to search for the account number, and specify the format it is in.

    Read the article

  • Push Email To Your BlackBerry

    If you are a BlackBerry user, then you definitely wish to have Email synchronization on your phone. But sometimes the process may seem difficult to new BlackBerry users and those who are unfamiliar w... [Author: Susan Brown - Computers and Internet - May 08, 2010]

    Read the article

  • Essbase Analytics Link (EAL) - Performance of some operation of EAL could be improved by tuning of EAL Data Synchronization Server (DSS) parameters

    - by Ahmed Awan
    Generally, performance of some operation of EAL (Essbase Analytics Link) could be improved by tuning of EAL Data Synchronization Server (DSS) parameters. a. Expected that DSS machine will be 64-bit machine with 4-8 cores and 5-8 GB of RAM dedicated to DSS. b. To change DSS configuration - open EAL Configuration Tool on DSS machine.     ->Next:     and define: "Job Units" as <Number of Cores dedicated to DSS> * 1.5 "Max Memory Size" (if this is 64-bit machine) - ~1G for each Job Unit. If DSS machine is 32-bit - max memory size is 2600 MB. "Data Store Size" - depends on number of bridges and volume of HFM applications, but in most cases 50000 MB is enough. This volume should be available in defined "Data Store Dir" driver.   Continue with configuration and finish it. After that, DSS should be restarted to take new definitions.  

    Read the article

  • MS Access 2010 hide/show text with abutton

    - by grant
    Hi I have a problem where the user has a form in MSAccess. The form contains information about the client. The client fields are – client’s first name, client’s last name, their street address, suburb and city, their landline number and their cell phone number and their email address. However the user does not always want to see the email address and would like to have a button that will show or hide the email address. I have to write a set of instructions that will solve this problem. Can anyone help??

    Read the article

  • Where can I find this HP modular keyboard or similar?

    - by Chaos Whistler
    I'm looking for an HP USB 104-Key Modular Keyboard with Detachable Number Pad (model number is DT530A, I believe), shown at this online store, where it's quite cheap but is sadly sold out. I've searched for hours on Google, and I've found several - the problem is they're Spanish or Greek keyboards but I need one in English (US keys preferred)! I would also be happy with a keyboard which is not exactly this but is similar, either with a detachable keypad or with no keypad at all, as long as the keyboard key layout and positioning is the same as on a standard keyboard. Can you help me find such a keyboard?

    Read the article

  • How do laptop battery voltages affect runtime?

    - by Bigbio2002
    I ordered a new battery for my faithful XPS M1710. I'm not sure of the voltage of the battery I have now, but the new one that the Dell rep got me (after 3-4 times confirming my phone number and laptop model number) is 14.8v. I was a bit concerned about potential incompatibilities (as most of the other compatible batteries listed were 11.1v), but I figure that there's no way that Dell would "recommend" batteries that wouldn't work or fry your system. Now, my question is, how does voltage affect battery life? If we assume the needed power draw to be constant, a higher voltage would indicate less amperage needed, therefore the battery would last longer before running out, yes? Or am I missing something? For reference: P=I*V P = power I = current V = voltage (duh)

    Read the article

  • How To Customize Your Android Lock Screen with WidgetLocker

    - by Jason Fitzpatrick
    Wouldn’t it be great to use your Android lock screen to turn on your flashlight, jump to your camera, and otherwise make accessing your phone and information on it lightening fast? Read on as we show you how. How To Customize Your Android Lock Screen with WidgetLocker The Best Free Portable Apps for Your Flash Drive Toolkit How to Own Your Own Website (Even If You Can’t Build One) Pt 3

    Read the article

  • joining text files with 600M+ lines

    - by dnkb
    I have two files huge.txt and small.txt. Huge has around 600M rows and it's 14Gigs, each line has four space separated words (tokens) and finally another space separated column with a number. Small has 150K rows with a size of ~3M, a space separated word and a number. Both Files are sorted using the sort command, with no extra options. The words in both files may include apostrophes (') and dashes (-). The deisred output would contain all columns from the huge.txt and the second column (the number) from small txt where the first word of huge.txt and the first word of small.txt match. My attemtpts below failed miserably with the following error: cat huge.txt|join -o 1.1 1.2 1.3 1.4 2.2 - small.txt > output.txt join: memory exhausted What I suspect is that the sorting order isn't right somehow even though the files are pre-sorted using: sort -k1 huge.unsorted.txt > huge.txt sort -k1 small.unsorted.txt > small.txt Problems seem to appear around words that have apostrophes (') or dashes (-). I also tried dictinoary sorting using the -d option bumping into the same error at the end. I see two ways out of this but don't know how to implement any of them. 1) Any tips how to sort the files in a way that the join command considers them to be sorted properly? 2) I was thinking of calculating MD5 or some other hashes of the strings to get rid of the apostrophes and dashes, and do the sorting and joining with the hashes instead of the strings themselves an dat the "translate" back the hashes to strings, but leave the numbers intact at the end of the lines. Since there would be only 150K hashes it's not that bad. What would be a good way to calculate individual hashes for each of the strings? Some AWK magic? See file samples at the end. Thank you! sample of huge.txt had stirred me to 46 had stirred my corruption 57 had stirred old emotions 55 had stirred something in 69 had stirred something within 40 sample of small.txt caley 114881 calf 2757974 calfed 137861 calfee 71143 calflora 154624 calfskin 148347 calgary 9416465 calgon's 94846

    Read the article

  • Highlighting duplicate column-pair and counting the rows Excel

    - by pleasehelpme
    Given the data below, the column-pair with the same values for at least 4 consecutive rows should be highlighted. image here for better visualization: http://i49.tinypic.com/2jeshtt.jpg 2 2 3 4 3 4 3 4 3 4 2 3 1 2 2 2 3 3 3 3 3 3 3 3 2 3 2 3 2 3 2 3 2 2 3 4 3 4 3 4 3 4 3 4 The output should be something like this, where the column-pair values that are the same for at least 4 consecutive rows are highlighted. image here for better visualization: http://i48.tinypic.com/i2lzc8.jpg 2 2 3 4 3 4 3 4 3 4 2 3 1 2 2 2 3 3 3 3 3 3 3 3 2 3 2 3 2 3 2 3 2 2 3 4 3 4 3 4 3 4 3 4 Then, I need to know the number of instances of the N-consecutive equal column-pair. Considering the data above, N=4 should be 3 and N=5 should be 1, where N is the number of rows that the column-pair is consecutively equal.

    Read the article

  • windows 2008r2 SNMP hrStorageSize reporting wrong values

    - by Giannis Nohj
    I'm trying to monitor disks on windows 2008 R2 system, but I'm getting wrong values returned from hrStorageSize and hrStorageUsed for one of the 2 disks the system has. Local disk C: is returning correct values, whereas D: disk, which is an SAN disk, is reporting 32G instead of 500G. Output of snmpwalk -v 2c -c public hostname-01 .1.3.6.1.2.1.25.2 : HOST-RESOURCES-MIB::hrStorageDescr.1 = STRING: C:\ Label: Serial Number 7ff87ca6 HOST-RESOURCES-MIB::hrStorageDescr.2 = STRING: D:\ Label:Data Serial Number dab5b0b9 ... HOST-RESOURCES-MIB::hrStorageAllocationUnits.1 = INTEGER: 4096 Bytes HOST-RESOURCES-MIB::hrStorageAllocationUnits.2 = INTEGER: 4096 Bytes ... HOST-RESOURCES-MIB::hrStorageSize.1 = INTEGER: 35834383 HOST-RESOURCES-MIB::hrStorageSize.2 = INTEGER: 7864320 ... HOST-RESOURCES-MIB::hrStorageUsed.1 = INTEGER: 32460169 HOST-RESOURCES-MIB::hrStorageUsed.2 = INTEGER: 3192302 I have calculated the first disk (C:) size as 136G, which is correct. On the other hand, D: is calculated as 30G, where it should be 500G. I also tried fsutil on the system and the output was correct. Anyone have any ideas?

    Read the article

  • Week in Geek: Google Announces New Round of Services to be Shut Down

    - by Asian Angel
    Our latest edition of WIG is filled with news link coverage on topics such as an IE flaw allows attackers and advertisers to track cursor movement, Microsoft will retire its Live Mesh PC-sync service in February, Yahoo has revamped its e-mail service & continues overhaul on Flickr, and more. Secure Yourself by Using Two-Step Verification on These 16 Web Services How to Fix a Stuck Pixel on an LCD Monitor How to Factory Reset Your Android Phone or Tablet When It Won’t Boot

    Read the article

  • How to schedule a biweekly cronjob?

    - by Roman
    crontab(5) defines the following fields: field allowed values ----- -------------- minute 0-59 hour 0-23 day of month 1-31 month 1-12 (or names, see below) day of week 0-7 (0 or 7 is Sun, or use names) and explains: Step values can be used in conjunction with ranges. Following a range with ``/<number>'' specifies skips of the number's value through the range. For example, ``0-23/2'' can be used in the hours field to specify command execution every other hour (the alternative in the V7 standard is ``0,2,4,6,8,10,12,14,16,18,20,22''). So, no biweekly Jobs, as far as my understanding goes. I'm quite sure there are workarounds, what are yours? Or did I miss something?

    Read the article

  • Reverse Proxy (mod_rewrite) and Rails (absolute paths)

    - by SooDesuNe
    I have front end rails app, that reverse proxies to any of a number of backend rails apps depending on URL, for example http://www.my_host.com/app_one reverse proxies to http://www.remote_host_running_app_one.com such that a URL like http://www.my_host.com/app_one/users will display the contents of http://www.remote_host_running_app_one.com/users I have a large, and ever expanding number of backends, so they can not be explicitly listed anywhere other than a database. This is no problem for mod_rewrite using a prg:/ rewrite map reverse proxy. The question is, the urls returned by rails helpers have the form /controller/action making them absolute to the root. This is a problem for the page served by mod_rewrite because links on the proxied page appear as absolute to the domain. i.e.: http://www.my_host.com/app_one/controller/action has links that end up looking like /controller/action/ when they need to look like /app_one/controller/action mod_proxy_html seems like the right idea, but it doesn't seem to be as dynamic as I would need, since the rules need to be hard coded into the config files. Is there a way to fix this server-side, so that the links will be routed correctly?

    Read the article

  • if a usb pen drive has 64349 cylinders, is it damaged?

    - by Andrew S
    I have a 4Gb USB pen drive that I'm trying to format with FAT32. When I run fdisk, it gives me this message: The number of cylinders for this disk is set to 64349. There is nothing wrong with that, but this is larger than 1024, and could in certain setups cause problems with: 1) software that runs at boot time (e.g., old versions of LILO) 2) booting and partitioning software from other OSs (e.g., DOS FDISK, OS/2 FDISK) I've tried deleting the partition table, creating new partitions, etc, but they never work. Sometimes I can write to the drive the first time I use it after formatting it, but then it becomes read-only in both Windows and Linux. I've tried this on multiple computers. Am I doing something wrong, or is the drive reporting an incorrect number of cylinders? Is the drive itself likely to be corrupted, and is there anyway to fix this under Windows Vista or Linux? Thanks

    Read the article

  • Dealing with Fine-Grained Cache Entries in Coherence

    - by jpurdy
    On occasion we have seen significant memory overhead when using very small cache entries. Consider the case where there is a small key (say a synthetic key stored in a long) and a small value (perhaps a number or short string). With most backing maps, each cache entry will require an instance of Map.Entry, and in the case of a LocalCache backing map (used for expiry and eviction), there is additional metadata stored (such as last access time). Given the size of this data (usually a few dozen bytes) and the granularity of Java memory allocation (often a minimum of 32 bytes per object, depending on the specific JVM implementation), it is easily possible to end up with the case where the cache entry appears to be a couple dozen bytes but ends up occupying several hundred bytes of actual heap, resulting in anywhere from a 5x to 10x increase in stated memory requirements. In most cases, this increase applies to only a few small NamedCaches, and is inconsequential -- but in some cases it might apply to one or more very large NamedCaches, in which case it may dominate memory sizing calculations. Ultimately, the requirement is to avoid the per-entry overhead, which can be done either at the application level by grouping multiple logical entries into single cache entries, or at the backing map level, again by combining multiple entries into a smaller number of larger heap objects. At the application level, it may be possible to combine objects based on parent-child or sibling relationships (basically the same requirements that would apply to using partition affinity). If there is no natural relationship, it may still be possible to combine objects, effectively using a Coherence NamedCache as a "map of maps". This forces the application to first find a collection of objects (by performing a partial hash) and then to look within that collection for the desired object. This is most naturally implemented as a collection of entry processors to avoid pulling unnecessary data back to the client (and also to encapsulate that logic within a service layer). At the backing map level, the NIO storage option keeps keys on heap, and so has limited benefit for this situation. The Elastic Data features of Coherence naturally combine entries into larger heap objects, with the caveat that only data -- and not indexes -- can be stored in Elastic Data.

    Read the article

  • Convert SWF/FLA back to PDF [closed]

    - by mitjak
    Possible Duplicate: How do I convert SWF into a PDF? I have a number of graphics and text only SWF and FLA files that I'd like to convert to PDF preserving the text and formatting. Printing to PDF from Flash Player seems to flatten everything into an image without preserving the text. I've got a large number of those files, so it would be quite nice to have a automatable/console utility to do the job. Appreciate any help or pointers. Thanks! EDIT: It seems like most SWF files were created with pdf2swf, so I'm hoping there is a similar tool for doing a reverse conversion.

    Read the article

  • Windows DFS Limitations

    - by Phil
    So far I have seen an article on performance and scalability mainly focusing on how long it takes to add new links. But is there any information about limitations regarding number of files, number of folders, total size, etc? Right now I have a single file server with millions of JPGs (approx 45 TB worth) that are shared on the network through several standard file shares. I plan to create a DFS namespace and replicate all these images to another server for high availability purposes. Will I encounter extra problems with DFS that I'm otherwise not experiencing with plain-jane file shares? Is there a more recommended way to replicate these millions of files and make them available on the network? EDIT: I would experiment on my own and write a blog post about it, but I don't have the hardware for the second server yet. I'd like to collect information before buying 45 TB of hard drive space...

    Read the article

  • headings numbering not updating

    - by Marwen Hizaoui
    I'm writing a report and I have some problems with headings. I have this structure: Chapter 1(heading1) 1.1(heading2) 1.1.1(heading3) 1.2(heading2) Chapter2(heading1) 2.1 2.2 The problem is that when I choose heading 1 word writes a number before chapter - 1 chapter 1 I want numbers to appear only from heading 2. I managed to change it using multilevel lists but it doesn't update appropriately. I mean I removed the number before chapter 1 for example but it did not update chapter 2 subheadings that became numbering from 1 (not 2). Please help me out. thanks.

    Read the article

  • rsync server side limit bandwidth/connection

    - by c2h2
    In a VOIP application, I have upto 3000 clients rsync audio files from there linux server in a daily, server is placed at a data center (10Mbps in/out bound), the server works as a VOIP sip server running FreeSWITCH (low ping latency should be ensured.) Therefore I would like to have server side control of rsync which controls: Limit total outbound bandwidth. Limit total number of connections. (Reject clients while at max number of connection and let it retry after a specific time frame.) OPTIONAL: list/kill individual connections. Normally I would use ssh + rsync + pem_keys with some extra options, but above requirements are not feasible by simple command lines. Can anyone point me some direction. or show some scripts/tools? I would also probably integrate them and release on github. Thanks!

    Read the article

  • Updating the jump in game

    - by Luka Tiger
    I am making a Java game and I want my game to run the same on any FPS so I'm using time delta between each update. This is the update method of the Player: public void update(long timeDelta) { //speed is the movement speed of a player on X axis //timeDelta is expressed in nano seconds so I'm dividing it with 1000000000 to express it in seconds if (Input.keyDown(37)) this.velocityX -= speed * (timeDelta / 1000000000.0); if (Input.keyDown(39)) this.velocityX += speed * (timeDelta / 1000000000.0); if(Input.keyPressed(38)) { this.velocityY -= 6; } velocityY += g * (timeDelta/1000000000.0); //applying gravity move(velocityX, velocityY); /*this is method which moves a player according to velocityX and velocityY, and checking the collision */ this.velocityX = 0.0; } The strange thing is that when I have unlimited FPS (and update number) my player is jumping about 10 blocks. It jumps even higher when the FPS is increasing. If I limit FPS it is jumping 4 blocks. (BLOCK: 32x32) I have just realized that the problem is this: if(Input.keyPressed(38)) { this.velocityY -= 6; } I add -6 to velocityY which increases player's Y proportionally to the update number and not to the time. But I don't know how to fix this.

    Read the article

  • Can I group rows to get sum using excel

    - by Matt
    I have a spreadsheet with 2 cols of importance. Date, and number. I can't always predict the number of rows or the date, but what I would like to do is print out the sum of the numbers for each date. For example, there might be 5 rows for Dec-7: 200, 111 and Dec-6: 222,533,100. I am tying to create a list which would show Dec-6: 855, Dec-7: 311. I believe a Pivot Table is what I want but I can't quite figure out how I need to configure it to show what I want. If anyone knows of a guide I could look at that would be fantastic!

    Read the article

  • How to See Your Estimated Data Usage in Windows 8

    - by Taylor Gibb
    Although you can use metered connections to get the most of your bandwidth in Windows 8, at times you may want to know how much data you have used for a single browsing session. Here’s how to do it. Why Your Android Phone Isn’t Getting Operating System Updates and What You Can Do About It How To Delete, Move, or Rename Locked Files in Windows HTG Explains: Why Screen Savers Are No Longer Necessary

    Read the article

  • Having the same texture data in different ID3D11Texture2D

    - by bdmnd
    Sorry if this has been answered elsewhere - I'm rather new to DX. My question concerns conservation of resources - specifically textures in VRAM. I assume that upon returning from a call to CreateTexture2D, a copy of any textures data supplied has been copied elsewhere, likely VRAM. Does DX11 have any facility for having multiple ID3D11Texture2D objects which point to the same data? This might at first seem silly, but imagine a ID3D11Texture2D which is an array of textures. In one material, an artist has chosen to blend three identically sized maps, saved on disk as A.dds, B.dds, and C.dds. Then imagine they have another material which also uses three maps, but this time A.dds, B.dds, and D.dds. The shader code knows the diffuse texture is a texture array, and also has the number of layers baked (three in each case). I would essentially like to set up just two ID3D11Texture2D objects, one for each material, but I don't want to waste VRAM for two identical copies of A.dds and B.dds. I could use explicit texture arrays, of course, but this reduces the number of resources available to the shader and can complicate code somewhat more than would otherwise be needed.

    Read the article

< Previous Page | 325 326 327 328 329 330 331 332 333 334 335 336  | Next Page >