Search Results

Search found 10279 results on 412 pages for 'video embedding'.

Page 134/412 | < Previous Page | 130 131 132 133 134 135 136 137 138 139 140 141  | Next Page >

  • How could I send live video stream to remote server from my phone !!!

    - by poc
    Hello , I have a problem about streaming my video to server in real-time from my phone. that is , let my phone be a IP Camera , and server can watch the live video from my phone I have googled many many solutions, but there is no one can solve my problem. I use MediaRecorder to record . it can save video file in the SD card correctly. then , I refered this page and used some method as followings skt = new Socket(InetAddress.getByName(hostname),port); pfd =ParcelFileDescriptor.fromSocket(skt); mediaRecorder.setOutputFile(pfd.getFileDescriptor()); now it seems I can send the video stream while recording however, I wrote a receiver-side program to receive the video stream from Android , but it doesn't work . is there any error? I can receive file , but I can not open the video file . I guess the problem may caused by file format ? there are outline of my code. in android side Socket skt = new Socket(hostIP,port); ParcelFileDescriptor pfd =ParcelFileDescriptor.fromSocket(skt); .... .... mediaRecorder.setAudioSource(MediaRecorder.AudioSource.MIC); mediaRecorder.setVideoSource(MediaRecorder.VideoSource.DEFAULT); mediaRecorder.setOutputFormat(MediaRecorder.OutputFormat.MPEG_4); mediaRecorder.setOutputFile(pfd.getFileDescriptor()); ..... mediaRecorder.setAudioEncoder(MediaRecorder.AudioEncoder.DEFAULT); mediaRecorder.setVideoEncoder(MediaRecorder.VideoEncoder.MPEG_4_SP); ..... mediaRecorder.start(); in receiver side (my ACER notebook) // anyway , I don't think the file extentions will do any effect File video = new File (strDate+".3gpp"); FileOutputStream fos; try { fos = new FileOutputStream(video); byte[] data = new byte[1024]; int count =-1; while( (count = fin.read(data,0,1024) ) !=-1) { fos.write(data,0,count); fos.flush(); } fos.close(); fin.close(); I confused a long time.... thanks in advance

    Read the article

  • Is this video card good for 3 monitors? [on hold]

    - by Brian Briu
    Video card: http://www.amazon.de/Zotac-NVIDIA-GeForce-Grafikkarte-Speicher/dp/B00D4KCSHG/ref=sr_1_11?ie=UTF8&qid=1408575042&sr=8-11&keywords=grafikkarte 3 Monitors like this: http://www.mediamarkt.de/mcs/product/ASUS-MX-239-H,48353,462790,951629.html?langId=-3 If yes how can I connect them? If no, i'll buy one more video card like that. Also tell me what motherboard should I buy for these things + 32 ram and i7 (remember that in the future i'll buy one morre video card (to have 2). Power supply also?

    Read the article

  • AxWindowsMediaPlayer does not play audio/video from url ?

    - by Madhup
    HI, I am using activeXMediaPlayer to play files from url but each time I pass a url to it shows the message , "either the file is corrupted or player does not support the file format u are playing." But when i run the same url on browser the file is downloaded and this downloaded file can be played on the media player. I am not able to find out what the problem is . Because the same cod plays the local file and the downloaded files but not file from url Although the same code worked few months ago for the urls So is this my fault or some server related issues can affect this thing. Please help me I am in big trouble. Regards, Madhup

    Read the article

  • iSightAudio.plugin error when playing video using MediaPlayer

    - by Elisabeth
    I am working on creating a simple iPhone app that plays a movie via URL. When I Build&Run to test in the simulator, it works fine; as soon as I start playing the movie, I get the following message in the console: [1757:4b03] Cannot find executable for CFBundle/CFPlugIn 0x820ffe0 </Library/Audio/Plug-Ins/HAL/iSightAudio.plugin> (not loaded) [1757:4b03] Cannot find function pointer iSightAudioNewPlugIn for factory 9BE7661E-8AEF-11D7-8692-000A959F49B0 in CFBundle/CFPlugIn 0x820ffe0 </Library/Audio/Plug-Ins/HAL/iSightAudio.plugin> (not loaded) I don't get this error on other programs, so I assume it has something to do with this specific program, which uses the MediaPlayer.framework. Does anyone know what is causing this problem and how to fix it? Thank you

    Read the article

  • Methods to Analyze Sell Orders (in a video game)

    - by Travis
    I'm building a market manager for EVE, but this problem should be solvable without any prior knowledge of the game. Basically, sell orders in EVE look like this: All I'm really concerned with is picking a good price. The numbers the program would pick are sent out to our production team so they can buy the items to build things. Just like in a real market, the numbers we send off need to reflect the cheapest price that still offers a good quantity. Also, available quantities change over time, so we can't just pick the lowest listed price. For example, in the screencap above, we would set the sell price between 2,800.00 ISK and 3,000.00 ISK, even if only 500 units were needed. Even though the cheapest order would fit the quantity, it would probably quickly disappear. It would be better to set the price a tad higher so we don't underpay the production people. The Question: It's easy enough to figure prices out just by looking at the data, but I can't figure out how to have a program [java] do it. Different items sell with much less or more quantities (i.e. 3 or 300 or 3000 items) so the program has to be able to modify itself based on the offerings. I don't really want code as much as I do a way to think this through. Thanks in advance.

    Read the article

  • iOS - playing a youtube video from a webview thumbnail

    - by soleil
    Was about to finally submit an app when I realized that the youtube videos I'm playing play from the iPhone, but not the iPad. Here's the code I'm using: NSString *embedHTML = @"\ <html><head>\ <style type=\"text/css\">\ body {\ background-color: transparent;\ color: white;\ }\ </style>\ </head><body style=\"margin:0\">\ <embed id=\"yt\" src=\"%@\" type=\"application/x-shockwave-flash\" \ width=\"%0.0f\" height=\"%0.0f\"></embed>\ </body></html>"; NSString *html = [NSString stringWithFormat:embedHTML, [videoItem objectForKey:@"URL"], thumb.frame.size.width, thumb.frame.size.height]; UIWebView *videoView = [[UIWebView alloc] initWithFrame:thumb.frame]; [videoView loadHTMLString:html baseURL:nil]; [videoContainer addSubview:videoView]; Is there any reason this would work on the iPhone and not the iPad? I'm testing on a 1st generation iPad. The videos play on both iPhone models I've tested (3GS and 4S). Nothing happens at all when I tap on the webviews on the iPad.

    Read the article

  • Building Website With Video Upload Feature - Possible Implementations?!

    - by Goober
    Hello! Currently I have a static html/css website that I want to build a backend onto in order to upload high quality videos to the site. I'm not entirely sure how I might go about doing this.......suggestions for implementations, language to use etc. would be greatly appreciated. I have no constraints aside from of course, spending millions of pounds......I plan to do this myself too.

    Read the article

  • Client / Server application for Video BroadCasting on LAN

    - by jbl
    I m much of a newbie to client/server applications and networking.I am currently starting up a project in which we are planning to create an application or something like a website accessible only to clients in that LAN network.We want the server to upload videos so that the clients can view them whenever they want.Also the clients can upload videos.Also we want some kind of registration to provide authentication to avoid bogus videos.Also additional features that can be added on later like comment section,ratings,Groups etc. Basically what we want is to build mini-youtube but with one major difference , we want the application to be LAN based. This is the idea behind the project.But we dont know where to start from as we are new to this stuff and also we have like 1 year to work on the project.Please guide me as to what all softwares i'll be needing to get this working !

    Read the article

  • Virtual PC on Windows 7 doesn't have adjustment for video card size?

    - by Jian Lin
    The current VirtualBox has a place where the video card size can be set by the user. It seems that Win 7's Virtual PC doesn't have one? Will it auto adjust -- but what if the screen size is 800 x 600 and the user resize it to 1600 x 1200, then the original video size may not be enough and will that cause any issue? I do sometimes see blinking random pixel region showing on the VPC's screen... maybe it is cause by not enough video RAM size?

    Read the article

  • photo and video sharing on facebook in asp.net

    - by IphoneDeveloper
    i have to share the photo on facebook for which i have written the following code try { labelSteps_1_2.Text = "<a name=\"fb_share\" type=\"button\"></a>" + "<script " + "src=\"http://static.ak.fbcdn.net/connect.php/js/FB.Share\" " + "type=\"text/javascript\"></script>"; HtmlMeta tag = new HtmlMeta(); tag.Name = "title"; tag.Content = "This is the Title"; Page.Header.Controls.Add(tag); HtmlLink link = new HtmlLink(); link.Href = "http://admin.bizofair.com/UploadOffers/O_O11090004_14032012053503_183_183.jpeg"; link.Attributes["rel"] = "image_src"; link.Style.Add("color", "red"); Page.Header.Controls.Add(link); } catch (Exception ex) { // Handle the exception } on page load but it is not showing the image and when i have paste url of my image in facebook debugger on facebook developer it is giving error Errors That Must Be Fixed Can't Download: Could not retrieve data from URL.

    Read the article

  • Generate FLV Video File with ActionScript

    - by ChPuK
    Hello. I'm working on Flash ActionScript. My Movie Load some data like photo, sounds and photo to the Movie Stage. Each of loaded MovieClip have own animation. Is it possible to generate (record) FLV or other Movie file from ActionScript with ActionScript?

    Read the article

  • Video/ Speech Development of Applications

    - by idea_
    Why do we continue to type and click away in IDEs when we could theoretically use hand gestures and speech to develop applications? Think about it - Developing a class by standing in-front of your computer, making some gesture, and yelling "CAR!". This doesn't have to strictly apply to OOP either. We have sufficient speech and image acquisition/ processing and analysis tools available to us, don't we? This seems plausible to me, but I may be overly ambitious. From a conceptual point-of-view, do you see any problems with the implementation?

    Read the article

  • Get frames from the video saved in the photo directory iphone?

    - by Ballu
    HI All. how can i get the frames of any video that i am fetching from the photo directory iphone .I checked out various links but they only provide any thumbnails or any particular TIME's frame image but i need the frames of full length video .And the respective audio too then is that possible that from the freames i got and the audio file i can create another video but will change the sequence of the frames only the audio file will be the same in the video.Is that feasible .? Sample code or link will be more helpful !!! Thanks, Balraj,

    Read the article

  • How to restrict http access to video files?

    - by Tharases
    I want to only let "right" people watch those videos. In other words, only registered users that are allowed (by other users, ie, friends) should see videos. I have a hard retriction for cpu usage in my shared environment, so I can use things like php's readfile.

    Read the article

  • How do I send user to youtube video on ListBoxItem click?

    - by Byron Sommardahl
    I'm creating a Windows Phone 7 app and learning Silverlight in the process. I have a youtube video url and I'd like the user to be able to click a video thumbnail in my list and go to a youtube app to view the video (similar to iPhone). Is this possible? If so, can you share some code to send the user to the video? I already have a MouseLeftButtonUp event handler.

    Read the article

  • Getting the DirectShow VideoRender filter to respond to MediaType changes on its Input Pin?

    - by Jonathan Websdale
    Below is the code extract from my decoder transform filter which takes in data from my source filter which is taking RTP network data from an IP camera. The source filter, decode filter can dynamically respond to changes in the camera image dimensions since I need to handle resolution changes in the decode library. I've used the 'ReceiveConnection' method as described in the DirectShow help, passing the new MediaType data in the next sample. However, I can't get the Video Mixing Renderer to accept the resolution changes dynamically even though the renderer will render the different resolution if the graph is stopped and restarted. Can anyone point out what I need to do to get the renderer to handle dynamic resolution changes? HRESULT CDecoder::Receive(IMediaSample* pIn) { //Input data does not necessarily correspond one-to-one //with output frames, so we must override Receive instead //of Transform. HRESULT hr = S_OK; //Deliver input to library long cBytes = pIn->GetActualDataLength(); BYTE* pSrc; pIn->GetPointer(&pSrc); try { hr = m_codec.Decode(pSrc, cBytes, (hr == S_OK)?&tStart : NULL); } catch (...) { hr = E_UNEXPECTED; } if (FAILED(hr)) { if (theLog.enabled()){theLog.strm() << "Decoder Error " << hex << hr << dec << " - resetting input"; theLog.write();} //Force reset of decoder m_bReset = true; m_codec.ResetInput(); //We have handled the error -- don't pass upstream or the source may stop. return S_OK; } //Extract and deliver any decoded frames hr = DeliverDecodedFrames(); return hr; } HRESULT CDecoder::DeliverDecodedFrames() { HRESULT hr = S_OK; for (;;) { DecodedFrame frame; bool bFrame = m_codec.GetDecodedFrame(frame); if (!bFrame) { break; } CMediaType mtIn; CMediaType mtOut; GetMediaType( PINDIR_INPUT, &mtIn); GetMediaType( PINDIR_OUTPUT, &mtOut); //Get the output pin's current image resolution VIDEOINFOHEADER* pvi = (VIDEOINFOHEADER*)mtOut.Format(); if( pvi->bmiHeader.biWidth != m_cxInput || pvi->bmiHeader.biHeight != m_cyInput) { HRESULT hr = GetPin(PINDIR_OUTPUT)->GetConnected()->ReceiveConnection(GetPin(PINDIR_OUTPUT), &mtIn); if(SUCCEEDED(hr)) { SetMediaType(PINDIR_OUTPUT, &mtIn); } } IMediaSamplePtr pOut; hr = m_pOutput->GetDeliveryBuffer(&pOut, 0, 0, NULL); if (FAILED(hr)) { break; } AM_MEDIA_TYPE* pmt; if (pOut->GetMediaType(&pmt) == S_OK) { CMediaType mt(*pmt); DeleteMediaType(pmt); SetMediaType(PINDIR_OUTPUT, &mt); pOut->SetMediaType(&mt); } // crop, tramslate and deliver BYTE* pDest; pOut->GetPointer(&pDest); m_pConverter->Convert(frame.Width(), frame.Height(), frame.GetY(), frame.GetU(), frame.GetV(), pDest); pOut->SetActualDataLength(m_pOutput->CurrentMediaType().GetSampleSize()); pOut->SetSyncPoint(true); if (frame.HasTimestamp()) { REFERENCE_TIME tStart = frame.Timestamp(); REFERENCE_TIME tStop = tStart+1; pOut->SetTime(&tStart, &tStop); } m_pOutput->Deliver(pOut); } return hr; }

    Read the article

  • How to reduce iOS AVPlayer start delay

    - by Bernt Habermeier
    Note, for the below question: All assets are local on the device -- no network streaming is taking place. The videos contain audio tracks. I'm working on an iOS application that requires playing video files with minimum delay to start the video clip in question. Unfortunately we do not know what specific video clip is next until we actually need to start it up. Specifically: When one video clip is playing, we will know what the next set of (roughly) 10 video clips are, but we don't know which one exactly, until it comes time to 'immediately' play the next clip. What I've done to look at actual start delays is to call addBoundaryTimeObserverForTimes on the video player, with a time period of one millisecond to see when the video actually started to play, and I take the difference of that time stamp with the first place in the code that indicates which asset to start playing. From what I've seen thus-far, I have found that using the combination of AVAsset loading, and then creating an AVPlayerItem from that once it's ready, and then waiting for AVPlayerStatusReadyToPlay before I call play, tends to take between 1 and 3 seconds to start the clip. I've since switched to what I think is roughly equivalent: calling [AVPlayerItem playerItemWithURL:] and waiting for AVPlayerItemStatusReadyToPlay to play. Roughly same performance. One thing I'm observing is that the first AVPlayer item load is slower than the rest. Seems one idea is to pre-flight the AVPlayer with a short / empty asset before trying to play the first video might be of good general practice. [http://stackoverflow.com/questions/900461/slow-start-for-avaudioplayer-the-first-time-a-sound-is-played] I'd love to get the video start times down as much as possible, and have some ideas of things to experiment with, but would like some guidance from anyone that might be able to help. Update: idea 7, below, as-implemented yields switching times of around 500 ms. This is an improvement, but it it'd be nice to get this even faster. Idea 1: Use N AVPlayers (won't work) Using ~ 10 AVPPlayer objects and start-and-pause all ~ 10 clips, and once we know which one we really need, switch to, and un-pause the correct AVPlayer, and start all over again for the next cycle. I don't think this works, because I've read there is roughly a limit of 4 active AVPlayer's in iOS. There was someone asking about this on StackOverflow here, and found out about the 4 AVPlayer limit: fast-switching-between-videos-using-avfoundation Idea 2: Use AVQueuePlayer (won't work) I don't believe that shoving 10 AVPlayerItems into an AVQueuePlayer would pre-load them all for seamless start. AVQueuePlayer is a queue, and I think it really only makes the next video in the queue ready for immediate playback. I don't know which one out of ~10 videos we do want to play back, until it's time to start that one. ios-avplayer-video-preloading Idea 3: Load, Play, and retain AVPlayerItems in background (not 100% sure yet -- but not looking good) I'm looking at if there is any benefit to load and play the first second of each video clip in the background (suppress video and audio output), and keep a reference to each AVPlayerItem, and when we know which item needs to be played for real, swap that one in, and swap the background AVPlayer with the active one. Rinse and Repeat. The theory would be that recently played AVPlayer/AVPlayerItem's may still hold some prepared resources which would make subsequent playback faster. So far, I have not seen benefits from this, but I might not have the AVPlayerLayer setup correctly for the background. I doubt this will really improve things from what I've seen. Idea 4: Use a different file format -- maybe one that is faster to load? I'm currently using .m4v's (video-MPEG4) H.264 format. I have not played around with other formats, but it may well be that some formats are faster to decode / get ready than others. Possible still using video-MPEG4 but with a different codec, or maybe quicktime? Maybe a lossless video format where decoding / setup is faster? Idea 5: Combination of lossless video format + AVQueuePlayer If there is a video format that is fast to load, but maybe where the file size is insane, one idea might be to pre-prepare the first 10 seconds of each video clip with a version that is boated but faster to load, but back that up with an asset that is encoded in H.264. Use an AVQueuePlayer, and add the first 10 seconds in the uncompressed file format, and follow that up with one that is in H.264 which gets up to 10 seconds of prepare/preload time. So I'd get 'the best' of both worlds: fast start times, but also benefits from a more compact format. Idea 6: Use a non-standard AVPlayer / write my own / use someone else's Given my needs, maybe I can't use AVPlayer, but have to resort to AVAssetReader, and decode the first few seconds (possibly write raw file to disk), and when it comes to playback, make use of the raw format to play it back fast. Seems like a huge project to me, and if I go about it in a naive way, it's unclear / unlikely to even work better. Each decoded and uncompressed video frame is 2.25 MB. Naively speaking -- if we go with ~ 30 fps for the video, I'd end up with ~60 MB/s read-from-disk requirement, which is probably impossible / pushing it. Obviously we'd have to do some level of image compression (perhaps native openGL/es compression formats via PVRTC)... but that's kind crazy. Maybe there is a library out there that I can use? Idea 7: Combine everything into a single movie asset, and seekToTime One idea that might be easier than some of the above, is to combine everything into a single movie, and use seekToTime. The thing is that we'd be jumping all around the place. Essentially random access into the movie. I think this may actually work out okay: avplayer-movie-playing-lag-in-ios5 Which approach do you think would be best? So far, I've not made that much progress in terms of reducing the lag.

    Read the article

< Previous Page | 130 131 132 133 134 135 136 137 138 139 140 141  | Next Page >