Search Results

Search found 11395 results on 456 pages for 'intel integrated graphics'.

Page 82/456 | < Previous Page | 78 79 80 81 82 83 84 85 86 87 88 89  | Next Page >

  • What is good RAM for my CPU?

    - by Jason94
    I'm going to upgrade my RAM, but I have no clue what to look for. There is cheap and there is expensive RAM. What should I look for? What are good attributes for intel i5? The i5 CPU and motherboeard supports 1600 MHz (PC3-12800), but there is still to much to choose from. There are CL10 (10-10-10-27) @ 1.5V CL11 @ 1.35V CL9 (9-9-9-24) @ 1.5V CL9 (9-9-9-24) @ 1.65V Higher CL is better? CL is classification? And I guess a high voltage is also good since thats what overclockers do to the ram (increase the voltage)?

    Read the article

  • determine the archetecture of a mac from the command line or script?

    - by Brian Postow
    I'm writing a shell script, and I need to know the archetecture, ie PPC or Intel. Back in the day, there was a program /bin/arch that told you, but my mac doesn't seem to have it... Is there an easy way I can do this? Grep for something in a logfile? call some other program that spits that out as a side effect? It would be nice to know what OS Version I'm running too, but that may not be necessary. thanks

    Read the article

  • 2x ATI FireMV 2250 not working

    - by Brett Rigby
    I've just bought 2x ATI FireMV 2250 cards from a reputable seller on ebay... one is a PCI-E x1 and the other is a PCI-E x16. Problem is, neither works in my Dell Vostro 220 PC... Both cards appear to not work. Is there something I'm missing? The seller used an anti-static bag for one of the cards, syaing that: "You Must connect this graphics board to the computer's power supply" but there's no socket for that, nor any indication. I have contacted the seller, but thought I'd raise it here, as it seems strange that two cards can be non-functioning.

    Read the article

  • Video driver reports 8-bit color depth after Windows Update

    - by user3163
    Windows Update pushed out a new video driver for the Mobile Intel(R) 4 Series Express Chipset Family. One HP 6730b laptop was updated from 8.15.10.1872 to 8.15.10.2021. After the update the color depth is reported as 8-bit even if the setting is 32-bit. This means that Areo effects are disabled. Downgrading to 8.15.10.1872 did not help. Is there any way to force the video driver to report 32-bit color depth?

    Read the article

  • Building my own computer

    - by There is nothing we can do
    I'm planning to build my own computer. I do not have enough cash to buy all components I need in one go. I want to ask, if I buy motherboard which is compatible with i7 processor (any) and compatible with graphic card Nvidia gtx 780, does it mean that this mother board will be compatible with processors (from intel) which will be released next year? Same for graphic cards? The point is that I'd like to avoid situation where I buy motherboard let's say now, and in couple of months there will be new graphic card/processor which will not be compatible with my mother board? Or maybe shall I start completely somewhere else?

    Read the article

  • Upgraded to a GTX 580 - no video signal on boot

    - by MGOwen
    I have: New Corsair 550w PSU (should be more than enough) AsRock H61M u3s3 mobo (with latest updated BIOS) Just replaced my faithful HD 7790 GPU with a refurbished EVGA GTX 580 1.5GB. When I boot, it sounds like the PC boots to windows, and the 580's fan is spinning, but I can't get any video signal from the card. Old card and integrated graphics still work fine (when I replace and use those instead). I am using the DVI ports on the card with a DVI-VGA adapter until I can get my hands on a mini HDMI adapter tomorrow. Anyone ever heard of a GPU needed something special to make the VGA signal work? Hold down a key on startup or something? I can't even find a manual for this card.

    Read the article

  • Does the size of the monitor Matter?

    - by Arsheep
    I have a old computer, and I want to buy a big LCD. The best I've found so far is Viewsonic's 24" LCD TFT monitor. So will it run without any problems, or do I need to upgrade the video cards or something as well? The computer is not too old: it has P4 board and celeron processor, with 128 graphics memory. And in display properties, it says that the maxium that I can use is 1280 x 1024 resolution. I am noob hardware-wise, so need help on this stuff. Thanks

    Read the article

  • Does the size of the monitor Matter?

    - by Arsheep
    I have a old computer, and I want to buy a big LCD. The best I've found so far is Viewsonic's 24" LCD TFT monitor. So will it run without any problems, or do I need to upgrade the video cards or something as well? The computer is not too old: it has P4 board and celeron processor, with 128 graphics memory. And in display properties, it says that the maxium that I can use is 1280 x 1024 resolution. I am noob hardware-wise, so need help on this stuff. Thanks

    Read the article

  • Does the size of the monitor Matter?

    - by Arsheep
    I have a old computer, and I want to buy a big LCD. The best I've found so far is Viewsonic's 24" LCD TFT monitor. So will it run without any problems, or do I need to upgrade the video cards or something as well? The computer is not too old: it has P4 board and celeron processor, with 128 graphics memory. And in display properties, it says that the maxium that I can use is 1280 x 1024 resolution. I am noob hardware-wise, so need help on this stuff. Thanks

    Read the article

  • Determine the architecture of a Mac from the command line or script?

    - by Brian Postow
    I'm writing a shell script, and I need to know the archetecture, ie PPC or Intel. Back in the day, there was a program /bin/arch that told you, but my mac doesn't seem to have it... Is there an easy way I can do this? Grep for something in a logfile? call some other program that spits that out as a side effect? It would be nice to know what OS Version I'm running too, but that may not be necessary. thanks

    Read the article

  • Does a dedicated video card improve HTML5 websites, Skype or Flash games performance?

    - by Kiewic
    I have read that having a dedicated video card (GPU) improves performance if you use your computer to play video games. I guess to make this happen, video games or apps must be using especial libraries designed to share the workload with the GPU, maybe Direct X or OpenGL, I don't know. Am I wrong? So, can HTML5 websites, Adobe Illustrator, Flash games (Zynga games), Skype or Netflix benefit from a dedicated video card? I usually do the previous activities simultaneously. Should I consider changing from an integrated video card to a dedicated card if I want to improve performance? Thanks.

    Read the article

  • How to makesure VM is in Para virtualization mode on XEN?

    - by ananthan
    I am having a server running XEN,with ubuntu 12.04 as Domain 0. Intel VT is turned on by default and i dont want to run any Windows OS as VM. So 1. how can i makesure that all the created VM will be in paravirtualized mode even with VT enabled? I have heard that in Full Virtualization hypervisor needs to work hard for emulating all the underlying hardware,which may reduce performance for too much i/o specific applications when compared to para mode. 2.Since i am running only Linux VMs do i really need to enable VT support? 3.Can i be able to run Linux VM in para mode with VT enabled? 4.Will simply turning on VT support changes virtualization from para to full virtualization?? 5.Will there be any performance improvement for VM running Linux OS if VT is enabled?

    Read the article

  • How come Core i7 (desktop) dominates Xeon (server)?

    - by grant tailor
    I have been using this performance benchmark results to select what CPUs to use on my web server and to my surprise, looks like Core i7 CPUs dominates the list pushing Xeon CPUs into the bush. Why is this? Why is Intel making the Core i7 perform better than the Xeon. Are Desktop CPUs supposed to perform better than server grade Xeon CPUs? I really don't get this and will like to know what you think or why this is so. Also I am thinking about getting a new web server and thinking between the i7-2600 VS the Xeon E3-1245. The i7-2600 is higher up in the performance benchmark but I am thinking the Xeon E3-1245 is server grade. What do you guys think? Should I go for the i7-2600? Or is the Xeon E3-1245 a server grade CPU for a reason?

    Read the article

  • How come i7 (desktop) dominates Xeon (server)?

    - by grant tailor
    I have been using this performance benchmark results http://www.cpubenchmark.net/high_end_cpus.html to select what CPUs to use on my web server and to my surprise...looks like i7 CPUs dominates the list pushing Xeon CPUs into the bush. Why is this? Why is Intel making the i7 perform better than the Xeon. Are Desktop CPUs supposed to perform better than server grade Xeon CPUs? I really don't get this and will like to know what you think or why this is so. Also i am thinking about getting a new web server and thinking between the i7-2600 VS the Xeon E3-1245. The i7-2600 is higher up in the performance benchmark but i am thinking the Xeon E3-1245 is server grade...so what do you guys think? Should i go for the i7-2600? Or is the Xeon E3-1245 a server grade CPU for a reason?

    Read the article

  • A GFX card with two DVI and one HDMI with audio?

    - by TomaszRykala
    I haven't upgraded my PC for a while nor kept up to date with pc hardware since 2007, but have recently come up with a requirement and I am wondering if I can solve it! Are there any graphics cards out there that have two DVIs and one HDMI output, with the HDMI supporting sound over it? Basically, I've got a two PC monitor setup, and both monitors are connected to the two DVI's of my GTX9800+. I want to upgrade my card to say, GTX 460, but can't find if there is such with two DVIs (for my monitors) and one HDMI with audio (for my TV). So I am wondering, if such solution possible at all? I noticed that there are some cards out there with two DVI's and one mini-HDMI, but they all say 'HDMI Output Yes x 1 (via DVI to HDMI adaptor x 1 )', which probably means that because of the DVI element, the audio won't work. Is this true? Such card is an exmaple: http://uk.asus.com/Graphics_Cards/NVIDIA_Series/ENGTX460_DirectCU2DI1GD5/#specifications Many thanks for your help!

    Read the article

  • What components should I upgrade to squeeze more life of my pc? [on hold]

    - by Jared
    Hi my current specs are CPU: Intel Core i5 750 2.66 GHz, Socket 1156 Motherboard: Asus P7P55 LX Motherboard, Socket 1156, 4xDIMM DDR3, 2xPCIe-16, 3xPCI, 2xPCIe-1, 14xUSB2, Audio, 1xATA, 6xSATA, RAID, ATX HDD: Seagate Barracuda 7200.11 ST31500341AS Hard Disk Drive, 1500GB, 7200rpm, 32MB Cache, SATA-2 RAM: G.Skill Ripjaw F3-12800CL8D-4GBRM, 2x2GB, DDR3-1600, PC3-12800, CL8, DIMM Video: MSI Radeon 5850 R5850-PM2D1G Video Card, 1024MB, DDR5, PCIe-16, DVI, CrossFire, HDMI PSU: Vantec ION2, 620W ATX PSU, SLI Ready, Black Firstly I know it badly needs an SSD, for which I'm thinking of the Samsung 250gb 840 EVO series. The ram could do with another 4GB, for which I am thinking of replacing the 2x2GB for 2x4GB of the same type. My video card is struggling with the latest releases, so vitally should I ditch the video card and go for something newer (ATI 7850, GTX 660?) or try and get another 5850 for crossfire support? Bearing in mind 5850's are sort of hard to come buy now as they are not stocked in shops usually, which means they have to be bought second hand.

    Read the article

  • Virtualizing Windows with VT-d to play within it

    - by Bahaïka
    I would like to know if what I want to is possible. First, here is my hardware configuration : MB : Gigabyte H77-DS3H Intel Core i5 3330 (3.00GHz) MSI Nvidia GTX 760 8 Gb RAM 1333 MHz What I want to do is to have a linux (probably ArchLinux) as a main and unique OS on my computer (I don't want a dualboot). And virtualizing Windows under Linux. I've read about VT-d (supported by my CPU) and PCIBackHide. I'd like to know if I can initialize Linux with the Integrated Graphics and Virtualize Windows to use the GTX 760. If I can do it, where do I have to plug my two screens ? On the MainBoard or the GTX 760 ? How do I achive that idea ?

    Read the article

  • A "quick" vector editor (SVG) for Linux (for annotating images?)

    - by sdaau
    I often need to take a bitmap (.png) image, and draw some lines or text on top of it, and possibly export a new, thusly "annotated" image. I know I can basically do all this in inkscape - but inkscape is a complex program, and it needs almost a minute to start up properly on my PCs. So I was thinking - is there something like a "mini" vector editor for Linux, which would start up fast, and allow me to: Right-click, open an image in this editor program The program scales the active "document"/"window size" to the size of the image I can zoom in/zoom out (and possibly crop) the image I can add at least lines, boxes and text in different colors? A bonus for me would be to have the overlay graphics saved as SVG format, say with the same filename as the image - as in, "image.png.svg" being saved in the same directory where the original "image.png" is located (thus allowing opening and editing these "annotations" further, either in this editor, or possibly in inkscape). And another bonus would be the export of the annotated image to a bitmap. Anyone know about anything like this?

    Read the article

  • Like to Upgrade My PC (7 year old) - for animation and hardcore gaming ! - help me [closed]

    - by sri
    I like to buy a new computer for my studies and as well as gaming. My old pc has 1.5 GB RAM with 512MB Graphics card. And it is very old to run Adobe CS5 version and other high end animation software. My budget will be INR 20k-25k. I have 500GB hard disk, keyboard and mouse new. So apart from this, I like to buy : Intel or AMD is good ? My idea is : Corei5 or Corei7 = which is best and economy ? Which mother board. 4 GB RAM with upto 8 GB RAM slot for future upgrade. 1 GB or 2 GB Graphic card - which one ? If I am wrong - please suggest me

    Read the article

  • SAML Identity Provider based on Active Directory

    - by Jarret
    I have a 3rd party program that supports web SSO using SAML 1.1 (it is ready to serve as the Service Provider, in other words). We would like to implement this SSO for our intranet users based on their Active Directory credentials. In other words, they've already logged on to their system, so let's simply use those credentials to facilitate an SSO. I am a little overwhelmed at where to begin, though. My initial thought is that IIS / Active Directory could easily serve as the Identity Provider since IIS gives us "Integrated Windows Authentication" abilities. I would think we could just create a .NET web app that requires Integrated Authentication which simply extracts the current user ID, builds the SAML response, and re-directs the user back to the Service Provider with this SAML response to complete the SSO. But then, my problem is that I simply have no real idea of how to go about creating this SAML response, the X.509 certs involved, etc... I am wondering if I am in over my head on this, or if creating this SAML response should be relatively easy. Note this SSO is to be used by intranet users only, so no need to worry about federating with other companies / domains.

    Read the article

  • App using MonoTouch Core Graphics mysteriously crashes

    - by Stephen Ashley
    My app launches with a view controller and a simple view consisting of a button and a subview. When the user touches the button, the subview is populated with scrollviews that display the column headers, row headers, and cells of a spreadsheet. To draw the cells, I use CGBitmapContext to draw the cells, generate an image, and then put the image into the imageview contained in the scrollview that displays the cells. When I run the app on the iPad, it displays the cells just fine, and the scrollview lets the user scroll around in the spreadsheet without any problems. If the user touches the button a second time, the spreadsheet redraws and continues to work perfectly, If, however, the user touches the button a third time, the app crashes. There is no exception information display in the Application Output window. My first thought was that the successive button pushes were using up all the available memory, so I overrode the DidReceiveMemoryWarning method in the view controller and used a breakpoint to confirm that this method was not getting called. My next thought was that the CGBitmapContext was not getting released and looked for a Monotouch equivalent of Objective C's CGContextRelease() function. The closest I could find was the CGBitmapContext instance method Dispose(), which I called, without solving the problem. In order to free up as much memory as possible (in case I was somehow running out of memory without tripping a warning), I tried forcing garbage collection each time I finished using a CGBitmapContext. This made the problem worse. Now the program would crash moments after displaying the spreadsheet the first time. This caused me to wonder whether the Garbage Collector was somehow collecting something necessary to the continued display of graphics on the screen. I would be grateful for any suggestions on further avenues to investigate for the cause of these crashes. I have included the source code for the SpreadsheetView class. The relevant method is DrawSpreadsheet(), which is called when the button is touched. Thank you for your assistance on this matter. Stephen Ashley public class SpreadsheetView : UIView { public ISpreadsheetMessenger spreadsheetMessenger = null; public UIScrollView cellsScrollView = null; public UIImageView cellsImageView = null; public SpreadsheetView(RectangleF frame) : base() { Frame = frame; BackgroundColor = Constants.backgroundBlack; AutosizesSubviews = true; } public void DrawSpreadsheet() { UInt16 RowHeaderWidth = spreadsheetMessenger.RowHeaderWidth; UInt16 RowHeaderHeight = spreadsheetMessenger.RowHeaderHeight; UInt16 RowCount = spreadsheetMessenger.RowCount; UInt16 ColumnHeaderWidth = spreadsheetMessenger.ColumnHeaderWidth; UInt16 ColumnHeaderHeight = spreadsheetMessenger.ColumnHeaderHeight; UInt16 ColumnCount = spreadsheetMessenger.ColumnCount; // Add the corner UIImageView cornerView = new UIImageView(new RectangleF(0f, 0f, RowHeaderWidth, ColumnHeaderHeight)); cornerView.BackgroundColor = Constants.headingColor; CGColorSpace cornerColorSpace = null; CGBitmapContext cornerContext = null; IntPtr buffer = Marshal.AllocHGlobal(RowHeaderWidth * ColumnHeaderHeight * 4); if (buffer == IntPtr.Zero) throw new OutOfMemoryException("Out of memory."); try { cornerColorSpace = CGColorSpace.CreateDeviceRGB(); cornerContext = new CGBitmapContext (buffer, RowHeaderWidth, ColumnHeaderHeight, 8, 4 * RowHeaderWidth, cornerColorSpace, CGImageAlphaInfo.PremultipliedFirst); cornerContext.SetFillColorWithColor(Constants.headingColor.CGColor); cornerContext.FillRect(new RectangleF(0f, 0f, RowHeaderWidth, ColumnHeaderHeight)); cornerView.Image = UIImage.FromImage(cornerContext.ToImage()); } finally { Marshal.FreeHGlobal(buffer); if (cornerContext != null) { cornerContext.Dispose(); cornerContext = null; } if (cornerColorSpace != null) { cornerColorSpace.Dispose(); cornerColorSpace = null; } } cornerView.Image = DrawBottomRightCorner(cornerView.Image); AddSubview(cornerView); // Add the cellsScrollView cellsScrollView = new UIScrollView (new RectangleF(RowHeaderWidth, ColumnHeaderHeight, Frame.Width - RowHeaderWidth, Frame.Height - ColumnHeaderHeight)); cellsScrollView.ContentSize = new SizeF (ColumnCount * ColumnHeaderWidth, RowCount * RowHeaderHeight); Size iContentSize = new Size((int)cellsScrollView.ContentSize.Width, (int)cellsScrollView.ContentSize.Height); cellsScrollView.BackgroundColor = UIColor.Black; AddSubview(cellsScrollView); CGColorSpace colorSpace = null; CGBitmapContext context = null; CGGradient gradient = null; UIImage image = null; int bytesPerRow = 4 * iContentSize.Width; int byteCount = bytesPerRow * iContentSize.Height; buffer = Marshal.AllocHGlobal(byteCount); if (buffer == IntPtr.Zero) throw new OutOfMemoryException("Out of memory."); try { colorSpace = CGColorSpace.CreateDeviceRGB(); context = new CGBitmapContext (buffer, iContentSize.Width, iContentSize.Height, 8, 4 * iContentSize.Width, colorSpace, CGImageAlphaInfo.PremultipliedFirst); float[] components = new float[] {.75f, .75f, .75f, 1f, .25f, .25f, .25f, 1f}; float[] locations = new float[]{0f, 1f}; gradient = new CGGradient(colorSpace, components, locations); PointF startPoint = new PointF(0f, (float)iContentSize.Height); PointF endPoint = new PointF((float)iContentSize.Width, 0f); context.DrawLinearGradient(gradient, startPoint, endPoint, 0); context.SetLineWidth(Constants.lineWidth); context.BeginPath(); for (UInt16 i = 1; i <= RowCount; i++) { context.MoveTo (0f, iContentSize.Height - i * RowHeaderHeight + (Constants.lineWidth/2)); context.AddLineToPoint((float)iContentSize.Width, iContentSize.Height - i * RowHeaderHeight + (Constants.lineWidth/2)); } for (UInt16 j = 1; j <= ColumnCount; j++) { context.MoveTo((float)j * ColumnHeaderWidth - Constants.lineWidth/2, (float)iContentSize.Height); context.AddLineToPoint((float)j * ColumnHeaderWidth - Constants.lineWidth/2, 0f); } context.StrokePath(); image = UIImage.FromImage(context.ToImage()); } finally { Marshal.FreeHGlobal(buffer); if (gradient != null) { gradient.Dispose(); gradient = null; } if (context != null) { context.Dispose(); context = null; } if (colorSpace != null) { colorSpace.Dispose(); colorSpace = null; } // GC.Collect(); //GC.WaitForPendingFinalizers(); } UIImage finalImage = ActivateCell(1, 1, image); finalImage = ActivateCell(0, 0, finalImage); cellsImageView = new UIImageView(finalImage); cellsImageView.Frame = new RectangleF(0f, 0f, iContentSize.Width, iContentSize.Height); cellsScrollView.AddSubview(cellsImageView); } private UIImage ActivateCell(UInt16 column, UInt16 row, UIImage backgroundImage) { UInt16 ColumnHeaderWidth = (UInt16)spreadsheetMessenger.ColumnHeaderWidth; UInt16 RowHeaderHeight = (UInt16)spreadsheetMessenger.RowHeaderHeight; CGColorSpace cellColorSpace = null; CGBitmapContext cellContext = null; UIImage cellImage = null; IntPtr buffer = Marshal.AllocHGlobal(4 * ColumnHeaderWidth * RowHeaderHeight); if (buffer == IntPtr.Zero) throw new OutOfMemoryException("Out of memory: ActivateCell()"); try { cellColorSpace = CGColorSpace.CreateDeviceRGB(); // Create a bitmap the size of a cell cellContext = new CGBitmapContext (buffer, ColumnHeaderWidth, RowHeaderHeight, 8, 4 * ColumnHeaderWidth, cellColorSpace, CGImageAlphaInfo.PremultipliedFirst); // Paint it white cellContext.SetFillColorWithColor(UIColor.White.CGColor); cellContext.FillRect(new RectangleF(0f, 0f, ColumnHeaderWidth, RowHeaderHeight)); // Convert it to an image cellImage = UIImage.FromImage(cellContext.ToImage()); } finally { Marshal.FreeHGlobal(buffer); if (cellContext != null) { cellContext.Dispose(); cellContext = null; } if (cellColorSpace != null) { cellColorSpace.Dispose(); cellColorSpace = null; } // GC.Collect(); //GC.WaitForPendingFinalizers(); } // Draw the border on the cell image cellImage = DrawBottomRightCorner(cellImage); CGColorSpace colorSpace = null; CGBitmapContext context = null; Size iContentSize = new Size((int)backgroundImage.Size.Width, (int)backgroundImage.Size.Height); buffer = Marshal.AllocHGlobal(4 * iContentSize.Width * iContentSize.Height); if (buffer == IntPtr.Zero) throw new OutOfMemoryException("Out of memory: ActivateCell()."); try { colorSpace = CGColorSpace.CreateDeviceRGB(); // Set up a bitmap context the size of the whole grid context = new CGBitmapContext (buffer, iContentSize.Width, iContentSize.Height, 8, 4 * iContentSize.Width, colorSpace, CGImageAlphaInfo.PremultipliedFirst); // Draw the original grid into the bitmap context.DrawImage(new RectangleF(0f, 0f, iContentSize.Width, iContentSize.Height), backgroundImage.CGImage); // Draw the cell image into the bitmap context.DrawImage(new RectangleF(column * ColumnHeaderWidth, iContentSize.Height - (row + 1) * RowHeaderHeight, ColumnHeaderWidth, RowHeaderHeight), cellImage.CGImage); // Convert the bitmap back to an image backgroundImage = UIImage.FromImage(context.ToImage()); } finally { Marshal.FreeHGlobal(buffer); if (context != null) { context.Dispose(); context = null; } if (colorSpace != null) { colorSpace.Dispose(); colorSpace = null; } // GC.Collect(); //GC.WaitForPendingFinalizers(); } return backgroundImage; } private UIImage DrawBottomRightCorner(UIImage image) { int width = (int)image.Size.Width; int height = (int)image.Size.Height; float lineWidth = Constants.lineWidth; CGColorSpace colorSpace = null; CGBitmapContext context = null; UIImage returnImage = null; IntPtr buffer = Marshal.AllocHGlobal(4 * width * height); if (buffer == IntPtr.Zero) throw new OutOfMemoryException("Out of memory: DrawBottomRightCorner()."); try { colorSpace = CGColorSpace.CreateDeviceRGB(); context = new CGBitmapContext (buffer, width, height, 8, 4 * width, colorSpace, CGImageAlphaInfo.PremultipliedFirst); context.DrawImage(new RectangleF(0f, 0f, width, height), image.CGImage); context.BeginPath(); context.MoveTo(0f, (int)(lineWidth/2f)); context.AddLineToPoint(width - (int)(lineWidth/2f), (int)(lineWidth/2f)); context.AddLineToPoint(width - (int)(lineWidth/2f), height); context.SetLineWidth(Constants.lineWidth); context.SetStrokeColorWithColor(UIColor.Black.CGColor); context.StrokePath(); returnImage = UIImage.FromImage(context.ToImage()); } finally { Marshal.FreeHGlobal(buffer); if (context != null){ context.Dispose(); context = null;} if (colorSpace != null){ colorSpace.Dispose(); colorSpace = null;} // GC.Collect(); //GC.WaitForPendingFinalizers(); } return returnImage; } }

    Read the article

  • Right way to dispose Image/Bitmap and PictureBox

    - by kornelijepetak
    I am trying to develop a Windows Mobile 6 (in WF/C#) application. There is only one form and on the form there is only a PictureBox object. On it I draw all desired controls or whatever I want. There are two things I am doing. Drawing custom shapes and loading bitmaps from .png files. The next line locks the file when loading (which is an undesired scenario): Bitmap bmp = new Bitmap("file.png"); So I am using another way to load bitmap. public static Bitmap LoadBitmap(string path) { using (Bitmap original = new Bitmap(path)) { return new Bitmap(original); } } This is I guess much slower, but I don't know any better way to load an image, while quickly releasing the file lock. Now, when drawing an image there is method that I use: public void Draw() { Bitmap bmp = new Bitmap(240,320); Graphics g = Graphics.FromImage(bmp); // draw something with Graphics here. g.Clear(Color.Black); g.DrawImage(Images.CloseIcon, 16, 48); g.DrawImage(Images.RefreshIcon, 46, 48); g.FillRectangle(new SolidBrush(Color.Black), 0, 100, 240, 103); pictureBox.Image = bmp; } This however seems to be some kind of a memory leak. And if I keep doing it for too long, the application eventually crashes. Therefor, I have X questions: 1.) What is the better way for loading bitmaps from files without locking the file? 2.) What objects needs to be manually disposed in the Draw() function (and in which order) so there's no memory leak and no ObjectDisposedException throwing? 3.) If pictureBox.Image is set to bmp, like in the last line of the code, would pictureBox.Image.Dispose() dispose only resources related to maintaining the pictureBox.Image or the underlying Bitmap set to it?

    Read the article

  • Getting a gestureoverlayview

    - by Codejoy
    I have been using some nice tutorials on drawing graphics on my android. I wanted to also add in the cool gesture demo found here: http://developer.android.com/resources/articles/gestures.html That takes these lines of code: GestureOverlayView gestures = (GestureOverlayView) findViewById(R.id.gestures); gestures.addOnGesturePerformedListener(this); This is fine and dandy yet I realize in my demo i'm trying to build using code from "Playing with Graphics in Android". The demos make sense, everything makes sense but I found out by using: setContentView(new Panel(this)); as is required by the Playing With Graphics tutorials, then the findViewById seems to no longer be valid and returns null. At first I was about to post a stupider question as to why this is happening, a quick test of playing with the setContentView made me realize the cause of findViewById returning null, I just do not know how to remedy this issue. Whats the key I am missing here? I realize that the new Panel is doinking some reference up but I am not sure how to make the connection here. THe: R.id.gestures is defined right int he main.xml as: (just like the tutorial) Thanks, Shane p.s. im new here be gentle.

    Read the article

  • Getting a gestureoverlayview

    - by Codejoy
    I have been using some nice tutorials on drawing graphics on my android. I wanted to also add in the cool gesture demo found here: http://developer.android.com/resources/articles/gestures.html That takes these lines of code: GestureOverlayView gestures = (GestureOverlayView) findViewById(R.id.gestures); gestures.addOnGesturePerformedListener(this); This is fine and dandy yet I realize in my demo i'm trying to build using code from "Playing with Graphics in Android". The demos make sense, everything makes sense but I found out by using: setContentView(new Panel(this)); as is required by the Playing With Graphics tutorials, then the findViewById seems to no longer be valid and returns null. At first I was about to post a stupider question as to why this is happening, a quick test of playing with the setContentView made me realize the cause of findViewById returning null, I just do not know how to remedy this issue. Whats the key I am missing here? I realize that the new Panel is doinking some reference up but I am not sure how to make the connection here. The: R.id.gestures is defined right int he main.xml as: (just like the tutorial) <android.gesture.GestureOverlayView android:id="@+id/gestures" android:layout_width="fill_parent" android:layout_height="0dip" android:layout_weight="1.0" /> So I did confirm the setContentView(new Panel(this)) is causing the issue. So I know the issue is that I have to figure out how to add the android.gesture.GestureOverlayView to the panel class somehow, I am just not sure how to go about this. After fighting with this I generally know what I need to do just now how to do it. I think I need either the equivalent of creating a panel in that main.xml OR figuring out how to build whats in main.xml for the gestures in code. I am close because I did this: GestureOverlayView gestures = new GestureOverlayView(this); which gets me a non null gestures now, unfortunately since I am not telling it to fill Parent anywhere I don't think its really showing up, so I am trying hard to figure out layout pa rams. Am I even on the right track?

    Read the article

  • Is it possible to integrate Computer Associates SiteMinder with SQLServer?

    - by Scott Weinstein
    We have a MS SQL BI stack which the standards group wants us to move to "WebSSO" which is based on Computer Associates SiteMinder/netegrity product. I figure integrating the web component won't be too hard, but we have users which connect to the Database directly - currently using Windows Authentication. Is is possible to itegrate Computer Associates SiteMinder with SQL Server? With SSAS? If so, how much effort is involved?

    Read the article

< Previous Page | 78 79 80 81 82 83 84 85 86 87 88 89  | Next Page >