Search Results

Search found 5213 results on 209 pages for 'integrated camera'.

Page 27/209 | < Previous Page | 23 24 25 26 27 28 29 30 31 32 33 34  | Next Page >

  • Xcode save image to camera roll

    - by user1470914
    I am new to Xcode (using 4.3) and am not sure how to save an image to the device's camera roll. All that I have done so far is an IBAction for the button to save the image (the button resides in the top right of the navigation bar. This tutorial (http://www.youtube.com/watch?v=sSdwjQUXA2Y) is EXACTLY what I would like to be able to do but the code is too blurry to read properly. Any help would be greatly appreciated.

    Read the article

  • Display issues on new OpenSUSE install

    - by user1319182
    I installed OpenSUSE 13.1 on my newly built PC, but the display is just horrible : the edges of the screen are missing. For example, I can't see all the top part, I can barely read the date and I see "ctivities" instead of "Activities". However, when I take a screenshot everything seems to be fine (the cursor doesn't appear though) the characters are sometimes too big and sometimes too small the cursor is huge and many other strange things. I took a few pictures I'm using an Intel integrated GPU (HD4400) and made all the possible updates with YaST. Any idea how I can fix this? Thanks

    Read the article

  • How do I use UIImagePickerController just to display the camera and not take a picture?

    - by Thomas
    Hello All: I'd like to know how to open the camera inside of a pre-defined frame (not the entire screen). When the view loads, I have a box, and inside it, I want to display what the camera sees. I don't want to snap a picture, just basically use the camera as a viewfinder. I have searched this site and have not yet found what I'm looking for. Please help. Thanks! Thomas Update 1: Here is what I have tried so far. 1.) I added UIImageView to my xib. 2.) Connect the following outlet to the UIImageView in IB IBOutlet UIImageView *cameraWindow; 3.) I put the following code in viewWillAppear -(void)viewWillAppear:(BOOL)animated { [super viewWillAppear:animated]; UIImagePickerController *picker = [[UIImagePickerController alloc] init]; picker.delegate = self; picker.sourceType = UIImagePickerControllerSourceTypeCamera; [self presentModalViewController:picker animated:YES]; NSLog(@"viewWillAppear ran"); } But this method does not run, as evident by the absence of NSLog statement from my console. Please help! Thanks, Thomas Update 2: OK I got it to run by putting the code in viewDidLoad but my camera still doesn't show up...any suggestions? Anyone....? I've been reading the UIImagePickerController class reference, but am kinda unsure how to make sense of it. I'm still learning iPhone, so it's a bit of a struggle. Please help! - (void)viewDidLoad { [super viewDidLoad]; // Create a bool variable "camera" and call isSourceTypeAvailable to see if camera exists on device BOOL camera = [UIImagePickerController isSourceTypeAvailable:UIImagePickerControllerSourceTypeCamera]; // If there is a camera, then display the world throught the viewfinder if(camera) { UIImagePickerController *picker = [[UIImagePickerController alloc] init]; // Since I'm not actually taking a picture, is a delegate function necessary? picker.delegate = self; picker.sourceType = UIImagePickerControllerSourceTypeCamera; [self presentModalViewController:picker animated:YES]; NSLog(@"Camera is available"); } // Otherwise, do nothing. else NSLog(@"No camera available"); } Thanks! Thomas Update 3: A-HA! Found this on the Apple Class Reference. Discussion The delegate receives notifications when the user picks an image or movie, or exits the picker interface. The delegate also decides when to dismiss the picker interface, so you must provide a delegate to use a picker. If this property is nil, the picker is dismissed immediately if you try to show it. Gonna play around with the delegate now. Then I'm going to read on wtf a delegate is. Backwards? Whatever :-p Update 4: The two delegate functions for the class are – imagePickerController:didFinishPickingMediaWithInfo: – imagePickerControllerDidCancel: and since I don't actually want to pick an image or give the user the option to cancel, I am just defining the methods. They should never run though....I think.

    Read the article

  • What I need to know for writing a Camera Component

    - by Delphawi
    I want to write a component that uses 2 webcams (1 integrated in my laptop , the other is a USB webcam) What do I need to know (or have) to build a component to deal with the cameras (capture , record , movement recognition , and other image and video processing) ? and how ? (with C++ or Delphi) I just need to know the concepts and main techniques , any good resources or source codes would be great :) Thank you .

    Read the article

  • Does running IIS7 in classic mode affect MVC output caching?

    - by Bob
    I have a need to run an application in classic mode for backwards compatibility with a specific application, and am trying to understand what kind of impact that will have on the performance of an MVC application that is running on the site. If we put a few static file maps (for .js, .css, .png, etc) above the ASP.NET wildcard map to reduce the amount of processing by the ASP.NET handler, will we be approaching the integrated mode in terms of performance? The thing i'm primarily concerned with is any effect this might have on output caching. I understand that integrated mode might (?) allow for the output cache to handle non ASP.NET content, but that isn't really a concern. We're more interested in ensuring that the MVC application has full use of the output cache. Empirically i've found that the two configurations operate on par when things go well, but if the page references resources that are not available, the integrated mode tends to fail much more quickly than the classic mode (e.g. 500 ms vs 10 seconds), reducing 'hang time' on the page load. Thanks for any feedback.

    Read the article

  • Couldn't upload files to Sharepoint site while passing through Squid Proxy

    - by Ecio
    Hi all, we have this issue: one of our employees is collaborating with a supplier and he needs to upload documents on a Sharepoint site hosted on the supplier's main site. In our environment we use Squid Proxy to allow people navigate on the net (we have NTLM authentication and users transparently authenticate while using IE and FF). It seems that this specific Sharepoint site is using Integrated Windows Authentication only, and according to some research on the net it seems that this can have troubles with proxies. More specifically, we have tried two Squid versions: with Squid 3.0 we are unable to login to the site (the browser loads an empty page) with Squid 2.7 (that supports "Connection Pinning") we are able to login into the site, move on the different sections BUT.. when we try to upload a file that is bigger than a couple of KiloBytes (i.e. 10KB) the browser loads an error page (i think it's a 401 unauthorized but i must verify it) we've tried changing a couple of Squid options (in 2.7), what we got is that when you try to upload the file you got an authentication box (just like the initial login) and it refuses to go on even if you enter the same authentication credentials. What's really strange is that when you try to upload a small file (i.e. a text or binary 1KB file) the upload succeeds. I initially thought that maybe there was something misconfigured on their Sharepoint site but I've tried also this site: www.xsolive.com (it's a sharepoint 2007 demo site) and I've experienced the same problem. Has any of you experienced such behaviour? Thanks! Of course we've suggested to the supplier to activate also Basic+SSL and we're waiting for their reply..

    Read the article

  • ATI Radeon HD7000 Series (Laptop) - Switch Mode Between ATI & Intel Integrated GPU. Stuck on Boot Screen On Intel GPU Selection Mode

    - by Monkey Drone
    Laptop Specs: HP Pavilion G6-2020SE GPUs 1) ATI HD7000 Series 2) Intel Integrated OS Installed: x) Ubuntu 12.04 (64 bit) i) ATI Graphics Card Drivers Installed From AMD website. Note: Graphics Card Drivers are Working Fine in 3D Mode. It runs a little Hot as it should since its a GPU. Observation) AMD Catalyst Control Centre Lets me Choose If I want to run the system in HIGH-END (ATI GPU) OR Intel Integrated (Better battery life) While I am on High End GPU Choice, Ubuntu works fine. Problem) But when I switch to Intel Mode in the AMD CCC and reboot the Machine. Ubuntu goes into 'Low Graphics Mode'. The problem is not that it goes into low graphics mode, it is completely expected since I am no longer using the ATI GPU but the integrated Intel GPU. Problem starts with the 'Selection' of the options. During that screen, I have no mouse on the screen (even tried plugging in an external USB mouse) & No Keyboard functionality. Thus I am left completely disabled to choose any option and load into Ubuntu. The Only thing I can do is switch to a terminal and enable ATI GPU through command-line and Ubuntu works Fine again. Is it a bug that there is no mouse/keyboard available to me during the startup of Ubuntu when its launched in Low-Graphics Mode? Any suggestions on how to pass through that? My palms are sweating as I write this down because the ATI GPU is really heating up my laptop. I dont want to boot into Windows or keep it around any longer than necessary. Please advise with help and directions. Sincerely, MonkeyD Edit1: The Answer by Celso has helped me switch to Intel, thus giving me sufficient battery power. Kudos to Celso. Now I can at least use my laptop for the time being without having it burn hair off my skin. I am still looking for answer to my original question of, 'why is lightdm not working properly when I switch to Intel GPU using ATI HD7000 series official drivers provided by AMD'.

    Read the article

  • Pros and cons of integrated vs. standalone tools [closed]

    - by eckes
    When it comes to version control, for example, there seem to be two different types of users: those using standalone VCS tools those using integrated tools from their IDE In my opinion, both have pros: Integrated tools do everything out of the IDE, no need to leave the environment you're used to ... Standalone tools usable for every type of project, not only for those associated with an IDE always behave the same (e.g. no difference like Eclipse-SVN-Client vs. AnkhSVN client for VS) ... I would be interested in your opinions and use cases.

    Read the article

  • Wpf multiple cameras

    - by melculetz
    Hello everybody, I wanted to create an WPF 3D application that allows the user to load a MeshGeometry3D and view it from 2 perspective (back and front) in the same time, in 2 different windows. First I thought of creating 2 cameras inside the same viewport, but it seams that the Viewport3D class allows a single camera. Then I tried to create 2 viewports, each with a different camera position, but I cannot add the same object in both viewports. Is there any way of seeing the same object through 2 different cameras in the same time?

    Read the article

  • Android: How to save a preview frame as jpeg image?

    - by niko
    Hi, I would like to save a preview frame as a jpeg image. I have tried to write the following code: public void onPreviewFrame(byte[] _data, Camera _camera) { if(settings.isRecording()) { Camera.Parameters params = _camera.getParameters(); params.setPictureFormat(PixelFormat.JPEG); _camera.setParameters(params); String path = "ImageDir" + frameCount; fileRW.setPath(path); fileRW.WriteToFile(_data); frameCount++; } } but it's not possible to open a saved file as a jpeg image. Does anyone know how to save preview frames as jpeg images? Thanks

    Read the article

  • Taking video from video camera and displaying it with MPMoviePlayerController IPhone SDK

    - by Daniel
    Has anyone tried taking a video from the camera and then using the video player provided to play it? When you take the video in portrait mode, sometimes the movie will play (when the player puts it in landscape mode) and when it puts it in portrait mode you cannot view the movie all you hear is sound,sometimes in landscape mode is flickers and does not play right, has anyone encountered this and found a way to fix it? My code to play the video looks like this: - (void)imagePickerController:(UIImagePickerController *)picker didFinishPickingMediaWithInfo:(NSDictionary *)info { NSURL *urls=[info objectForKey:@"UIImagePickerControllerMediaURL"] ; moviePlayer = [[MPMoviePlayerController alloc] initWithContentURL:[info urls]]; if (moviePlayer) { [moviePlayer play]; } } I checked settings on the docs nothing seems like it would fix this...Thanks

    Read the article

  • How do I render 3d model into directshow virtual camera output

    - by Mr Bell
    I want to provide a virtual webcam via DirectShow that will use the video feed from an existing camera running some tracking software against it to find the users face and then overlay a 3d model oriented just that it appears to move the users face. I am using a third party api to do the face tracking and thats working great. I get position and rotation data from that api. My question is whats the best way to render the 3d model and get into the video feed and out to direct show? I am using c++ on windows xp.

    Read the article

  • Augmented reality problem with multiple webs ervice.

    - by sujitjitu
    Hi, i am working on an Augmented Reality application.I am using android platform. i am retrieving data by using Handler concept from web service to show it above the camera preview. But my problem is that when i am making one web service call it is working fine but with multiple web service call the camera is getting crashed.It is working fine in the emulator but not working in the real device. I don't know what is the problem. I have tried everything not got the solution yet. Please help me its urgent. Thanks in advance..........

    Read the article

  • Weird in my UIImagePickerController with camera

    - by Chun
    Hi everyone: I met a problem on using camera by scratch on iphone. Other people's code just run fine on my 3GS. But my code doesn't. When I implemented a UIViewController controller, and I add the following code in viewdidload: UIImagePickerController *picker = [UIImagePickerController alloc] init]; picker.source = UIImagePickerControllerSourceTypeCamera; picker.delegate = self; //Previously added all the delegate properly [self presentModalViewController:picker animated:YES]; Nothing comes out. I checked in the debugger, it shows that the picker is allocated, but here is the major difference between my and success one. picker._imagepickerflag.visible = 0; //others show 1; picker.UINavigationController._containerView: 0x0 ; // others have value. Can anyone help me on this, is there something wrong? Thank you.

    Read the article

  • Android PixelFormat per device

    - by Tobias Domhan
    analogous to this thread: stackoverflow.com/questions/2093594/opengl-extensions-available-on-different-android-devices I would like to collect the different PixelFormats the android devices use. To find out you must do the following: Parameters camParams = camera.getParameters(); int format = camParams.getPreviewFormat(); Now you got to find the number in the following list: developer.android.com/reference/android/graphics/PixelFormat.html#A_8 How to generally open the camera is described here: developer.android.com/resources/samples/ApiDemos/src/com/example/android/apis/graphics/CameraPreview.html I'll have a start: device: T-mobile G1 / HTC Dream android: 1.6 (cyanogen mod) format: YCbCr_420_SP so what formats do your android phones use? thanks in advance :D

    Read the article

  • Displaying Current Pictures From A Camera

    - by Kap
    Hi Guys, I would like to develop an application that will receive pictures from a camera and maybe afterwards save it in a database. This is what i want to do: When the picture is taken it is send to the program immediately (or the program must read the current picture taken) then display it. I will take pictures of many people or things so anytime a picture is taken i want to see the current picture displayed in the program. I have googled if i can see an example application so that i know that it possible so that i can do mine from scratch. But couldn't find any so i am not sure if it is possible to do it in java. So guys am asking for guidelines how i can do it in java. I just need the steps then i will program everything myself. Thanks.

    Read the article

  • Need to open the SMS/MMS app and have an image pre-loaded into the text input in iPhone OS

    - by jmurphy
    Hello, I am trying to open the SMS/MMS app on the iPhone and have an image that is saved in the camera roll pre loaded into the text field. Is this possible? Here is the code I have tried: NSString *urlString = [NSString stringWithString:@"sms://asset/asset.JPG?id=1000000041&ext=JPG"]; NSString *escaped = [urlString stringByAddingPercentEscapesUsingEncoding:NSUTF8StringEncoding]; NSLog(@"URL: %@", escaped); [[UIApplication sharedApplication] openURL:[NSURL URLWithString:escaped]]; I received the URL from a return after I saved the image to the camera roll. Unfortunately this doesn't work. Any ideas? Thanks!

    Read the article

  • How to move a directional light according to the camera movement?

    - by Andrea Benedetti
    Given a light direction, how can I move it according to the camera movement, in a shader? Think that an artist has setup a scene (e.g., in 3DSMax) with a mesh in center of that and a directional light with a position and a target. From this position and target I've calculated the light direction. Now I want to use the same direction in my lighting equation but, obviously, I want that this light moves correctly with the camera. Thanks.

    Read the article

  • Intel lance un concours de création d'applications pour sa caméra 3D RealSense, avec plus d'un million de dollars US à se partager

    Intel lance un concours de création d'applications pour sa caméra 3D RealSense, avec plus d'un million de dollars US à se partager Mise à jour du 18/08/2014Intel annonce l'édition 2014 de son concours Perceptual Computing App Challenge articulé autour de sa technologie RealSense. Pour rappel, sa caméra 3D RealSense a été présentée comme étant son premier produit « d'informatique perceptuelle » et embarque entre autre un système de reconnaisse faciale, vocale, la suppression des arrières plans,...

    Read the article

  • Android Get Image Uri from Camera

    - by josnidhin
    Hi I have an application that calls the android phone's default camera to take photo the following is my code. Intent intent = new Intent(MediaStore.ACTION_IMAGE_CAPTURE); startActivityForResult(intent, TAKE_PICTURE); and in the onActivityResult method I am doing the following if ((requestCode == TAKE_PICTURE) && (resultCode == Activity.RESULT_OK)) { Uri photoPath = intent.getData(); // do something with the uri here } The above code works fine on htc Tatto and Sony ericsson's x10 running 1.6 but in on htc G1 running 1.6 the above code causes the following exception 03-08 18:54:25.906: ERROR/AndroidRuntime(4344): Uncaught handler: thread main exiting due to uncaught exception 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): java.lang.RuntimeException: Failure delivering result ResultInfo{who=null, request=1, result=-1, data=Intent { act=inline-data (has extras) }} to activity removed java.lang.NullPointerException 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread.deliverResults(ActivityThread.java:3224) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread.handleSendResult(ActivityThread.java:3266) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread.access$2600(ActivityThread.java:116) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1823) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.os.Handler.dispatchMessage(Handler.java:99) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.os.Looper.loop(Looper.java:123) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread.main(ActivityThread.java:4203) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at java.lang.reflect.Method.invokeNative(Native Method) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at java.lang.reflect.Method.invoke(Method.java:521) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:791) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:549) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at dalvik.system.NativeStart.main(Native Method) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): Caused by: java.lang.NullPointerException 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at removed 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.Activity.dispatchActivityResult(Activity.java:3624) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): at android.app.ActivityThread.deliverResults(ActivityThread.java:3220) 03-08 18:54:25.966: ERROR/AndroidRuntime(4344): ... 11 more Any insights into how to solve this problem. Thank you.

    Read the article

  • how to set the output image use com.android.camera.action.CROP

    - by adi.zean
    I have code to crop an image, like this : public void doCrop(){ Intent intent = new Intent("com.android.camera.action.CROP"); intent.setType("image/"); List<ResolveInfo> list = getPackageManager().queryIntentActivities(intent,0); int size = list.size(); if (size == 0 ){ Toast.makeText(this, "Cant find crop app").show(); return; } else{ intent.setData(selectImageUri); intent.putExtra("outputX", 300); intent.putExtra("outputY", 300); intent.putExtra("aspectX", 1); intent.putExtra("aspectY", 1); intent.putExtra("scale", true); intent.putExtra("return-data", true); if (size == 1) { Intent i = new Intent(intent); ResolveInfo res = list.get(0); i.setComponent(new ComponentName(res.activityInfo.packageName, res.activityInfo.name)); startActivityForResult(i, CROP_RESULT); } } } public void onActivityResult (int requestCode, int resultCode, Intent dara){ if (resultCode == RESULT_OK){ if (requestCode == CROP_RESULT){ Bundle extras = data.getExtras(); if (extras != null){ bmp = extras.getParcelable("data"); } File f = new File(selectImageUri.getPath()); if (f.exists()) f.delete(); Intent inten3 = new Intent(this, tabActivity.class); startActivity(inten3); } } } from what i have read, the code intent.putExtra("outputX", 300); intent.putExtra("outputY", 300); is use to set the resolution of crop result, but why i can't get the result image resolution higer than 300x300? when i set the intent.putExtra("outputX", 800); intent.putExtra("outputY", 800); the crop function has no result or crash, any idea for this situation? the log cat say "! ! ! FAILED BINDER TRANSACTION ! ! !

    Read the article

  • Creative Gesture Camera in Processing

    - by user2892963
    I'm trying to use the creative gesture camera in Processing. I started with the Intel Perceptual Computing SDK, and ran into an issue. I want to get the hand openness, and I am running into some issues - no matter what, the hand.openness returns 0. It otherwise runs quite well... Some Sample code I'm trying to get to work: If you open your hand it starts printing to the console, close it and it stops. import intel.pcsdk.*; PXCUPipeline session; PXCMGesture.GeoNode hand = new PXCMGesture.GeoNode(); void setup() { session = new PXCUPipeline(this); if(!session.Init(PXCUPipeline.GESTURE)) exit(); } void draw() { background(0); if(session.AcquireFrame(false)) { if(session.QueryGeoNode(PXCMGesture.GeoNode.LABEL_BODY_HAND_PRIMARY|PXCMGesture.GeoNode.LABEL_OPEN, hand)) //Only when primary hand is open { rect(0, 0, 10, 10); println(hand.openness + " : " + frameCount); //Openness should be from 0 to 100 } session.ReleaseFrame(); } } Using the current version of Processing (2.0.3), Perceptual Computing SDK Version 7383.

    Read the article

< Previous Page | 23 24 25 26 27 28 29 30 31 32 33 34  | Next Page >