Search Results

Search found 19871 results on 795 pages for 'commit log'.

Page 46/795 | < Previous Page | 42 43 44 45 46 47 48 49 50 51 52 53  | Next Page >

  • Get Mechanize to handle cookies from an arbitrary POST (to log into a website programmatically)

    - by Horace Loeb
    I want to log into https://www.t-mobile.com/ programmatically. My first idea was to use Mechanize to submit the login form: However, it turns out that this isn't even a real form. Instead, when you click "Log in" some javascript grabs the values of the fields, creates a new form dynamically, and submits it. "Log in" button HTML: <button onclick="handleLogin(); return false;" class="btnBlue" id="myTMobile-login"><span>Log in</span></button> The handleLogin() function: function handleLogin() { if (ValidateMsisdnPassword()) { // client-side form validation logic var a = document.createElement("FORM"); a.name = "form1"; a.method = "POST"; a.action = mytmoUrl; // defined elsewhere as https://my.t-mobile.com/Login/LoginController.aspx var c = document.createElement("INPUT"); c.type = "HIDDEN"; c.value = document.getElementById("myTMobile-phone").value; // the value of the phone number input field c.name = "txtMSISDN"; a.appendChild(c); var b = document.createElement("INPUT"); b.type = "HIDDEN"; b.value = document.getElementById("myTMobile-password").value; // the value of the password input field b.name = "txtPassword"; a.appendChild(b); document.body.appendChild(a); a.submit(); return true } else { return false } } I could simulate this form submission by POSTing the form data to https://my.t-mobile.com/Login/LoginController.aspx with Net::HTTP#post_form, but I don't know how to get the resultant cookie into Mechanize so I can continue to scrape the UI available when I'm logged in. Any ideas?

    Read the article

  • Using maven-release-plugin to tag and commit to non-origin

    - by Ali G
    When I do a release of my project, I want to share the source with a wider group of people than I normally do during development. The code is shared via a Git repository. To do this, I have used the following: remote public repository - released code is pushed here, every week or so (http://example.com/public) remote private repository - non-release code is pushed here, more than daily (http://example.com/private) In my local git repository, I have the following remotes defined: origin http://example.com/private public http://example.com/public I am currently trying to configure the maven-release-plugin to manage versioning of the builds, and to manage tagging and pushing of code to the public repository. In my pom.xml, I have listed the <scm/ as follows: <scm><connection>scm:git:http://example.com/public</connection></scm> (Removing this line will cause mvn release:prepare to fail) However, when calling mvn release:clean release:prepare release:perform Maven calls git push origin tagname rather than pushing to the URL specified in the POM. So the questions are: Best practice: Should I just be tagging and committing in my private repo (origin), and pushing to public manually? Can I make Maven push to the repository that I choose, rather than defaulting to origin? I felt this was implied by the requirement of the <connection/ element in <scm/.

    Read the article

  • Archive log transfer from Oracle 9i to Oracle 10g

    - by Jamie Love
    Hi all, I have a situation where I need to transfer Oracle 9i archive logs to an Oracle 10g database, from where they are to be mined by a log-miner and then used by an Oracle streams capture/apply processes. (Oracle 9 archive logs can be read by the Oracle 10 logminer - I can manually copy the archive logs across, manually register them and have them mined, captured then applied). The difficulty is that the way Oracle does archive log transfer changed quite a bit between 9i and 10g and setting up the 9i database to transfer to the remote machine like so: log_archive_dest_state_2 = enable log_archive_dest_2 = "service=OTHERMACHINE arch optional" no longer works. I get this in the 9i logs: *** 2009-05-22 04:03:44.149 RFS network connection lost at host 'OTHERMACHINE' Error 3113 attaching RFS server to standby instance at host 'OTHERMACHINE' Error 3113 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'OTHERMACHINE' Heartbeat failed to connect to standby 'OTHERMACHINE'. Error is 3113. *** 2009-05-22 04:03:44.150 kcrrfail: dest:2 err:3113 force:0 ORA-03113: end-of-file on communication channel And in the 10g log I get: Fri May 22 04:07:42 2009 WARNING: inbound connection timed out (ORA-3136) My question is: Does anyone know how I could configure my 9i or 10g server such that the 10g server will accept the 9i connection in such a way that I can transfer the 9i archive logs to the 10g server. It would be a bonus if the archive logs would be automatically registered in the 10g server. Note I have not set up a full DataGuard configuration here and the 10g database is not a secondary server. Thanks for any suggestions. Edit Note that I can log on to the 10g server from the 9i server via sqlplus, so connectivity is not the problem Edit 2 After a large amount of time searching for a solution, I've finally decided that such a mechanism doesn't work, and that a non-Oracle method of transferring archive logs from 9i to 10g will need to be used (e.g. rsync).

    Read the article

  • IIS7 Windows Server 2008 FTP -> Response: 530 User cannot log in

    - by RSolberg
    I just launched my first IIS FTP site following many of the tutorials from IIS.NET... I'm using IIS Users and Permissions rather than anonymous and/or basic. This is what I'm seeing while trying to establish the connection... Status: Resolving address of ftp.mydomain.com Status: Connecting to ###.###.##.###:21... Status: Connection established, waiting for welcome message... Response: 220 Microsoft FTP Service Command: USER MyFTPUser Response: 331 Password required for MyFTPUser. Command: PASS ******************** Response: 530 User cannot log in. Error: Critical error Error: Could not connect to server

    Read the article

  • Log internet requests made by vpn client

    - by Adam Schiavone
    How can I log the requests that people make to the internet when they are connected to my VPN? If one of them does something illegal, I need to be able to prove that it was not me, but a VPN client. This really is a just in case type thing. So, I want to be able to see the details (server, port, domain name, and time connected) in a file somewhere. Does windows offer this with its PPTP VPN service, or is there 3rd party software to do this (Free is better)? EDIT I have setup accounting, but that does not seem to tell me what I am looking for. Could someone comment on what accounting does? EDIT 2 Maybe I should heve been clearer. My system is running windows server 2008 r2 and using the stock VPN software (PPTP)

    Read the article

  • Understanding mongod log entries

    - by Jo Erlang
    Can anyone explain the numbers in the % column, in this mongod log snippet ? It is building a new index but I'm not sure I understand why the numbers look like that Tue Nov 20 11:38:08 [initandlisten] connection accepted from 127.0.0.1:49299 #8 (2 connections now open) Tue Nov 20 11:38:08 [conn8] build index xx.yyyy { source: 1 } Tue Nov 20 11:38:19 [conn8] 2921300/243339 1200% Tue Nov 20 11:38:29 [conn8] 4109600/243339 1688% Tue Nov 20 11:38:39 [conn8] 4400100/243339 1808% Tue Nov 20 11:38:49 [conn8] 4676600/243339 1921% Tue Nov 20 11:38:59 [conn8] 4939700/243339 2029% Tue Nov 20 11:39:09 [conn8] 5217800/243339 2144% Tue Nov 20 11:39:19 [conn8] 5439300/243339 2235% Tue Nov 20 11:39:29 [conn8] 5659700/243339 2325% Thanks

    Read the article

  • Help make sense of a KillDisk error/log

    - by user284194
    I have a hard drive that I've been trying to reformat. I tried reformatting it in the windows XP and 7 installers, and in an Ubuntu live cd with gparted. I tried using dd to 'zero' the drive as well with no success. Finally I ran across KillDisk after a search. I tried to zero the disk again with KillDisk and after 8 hours of zeroing I get the following errors in the log: ----------------------------------------Erase Session Begin--------------------------------------- 2010-03-23 19:35:54 Active@ KILLDISK for Windows Build 5.1.39 started Target: WDC WD2500KS-00MJB0 232.9 GB Located on: WDC WD2500KS-00MJB0 (Serial number: WD-WCANK9604799) Erase method: One Pass Zeros (1 pass) Passes: 1 Bad (unwritable) sectors detected from 1701 to 488397167 on Hard Disk 1. Error (the handle is invalid) refreshing device Hard Disk 1. Error (the handle is invalid) reading sector 0 on 81h. 2010-03-24 02:28:25 Total number of erased device(s): 0, partition(s): 0 -----------------------------------------Erase Session End---------------------------------------- Is the drive dead?

    Read the article

  • Nginx access log shows authenticated user "admin"

    - by bearcat
    I came across a line in my Nginx access log: 218.201.121.99 - admin [12/Dec/2012:18:33:18 +0800] "GET /manager/html HTTP/1.1" 444 0 "-" "-" Let me stress that there is only 1 record with this IP. Notice the authenticated user admin. After some googling, I was able to find out only that this is authenticated user (http://wiki.nginx.org/HttpCoreModule#.24remote_user), which was authenticated by the Auth Basic Module (http://wiki.nginx.org/HttpAuthBasicModule). However, nowhere in my site (configuration) do I use HTTP basic authentication. What is going on? How did it get there? Was the user authenticated?

    Read the article

  • Ibatis startBatch() only works with SqlMapClient's own start and commit transactions, not with Sprin

    - by Brian
    Hi, I'm finding that even though I have code wrapped by Spring transactions, and it commits/rolls back when I would expect, in order to make use of JDBC batching when using Ibatis and Spring I need to use explicit SqlMapClient transaction methods. I.e. this does batching as I'd expect: dao.getSqlMapClient().startTransaction(); dao.getSqlMapClient().startBatch(); int i = 0; for (MyObject obj : allObjects) { dao.storeChange(obj); i++; if (i % DB_BATCH_SIZE == 0) { dao.getSqlMapClient().executeBatch(); dao.getSqlMapClient().startBatch(); } } dao.getSqlMapClient().executeBatch(); dao.getSqlMapClient().commitTransaction(); but if I don't have the opening and closing transaction statements, and rely on Spring to manage things (which is what I want to do!), batching just doesn't happen. Given that Spring does otherwise seem to be handling its side of the bargain regarding transaction management, can anyone advise on any known issues here? (Database is MySQL; I'm aware of the issues regarding its JDBC pseudo-batch approach with INSERT statement rewriting, that's definitely not an issue here)

    Read the article

  • Log4j Relative Path to Tomcat7 Log Directory File Not Found

    - by Nik
    I have this log4j XML file: <log4j:configuration xmlns:log4j="http://jakarta.apache.org/log4j/"> <appender name="gemFile" class="org.apache.log4j.RollingFileAppender"> <param name="File" value="file:${catalina.home}/logs/GEM.log" /> <param name="MaxBackupIndex" value="5" /> <param name="MaxFileSize" value="3MB" /> <param name="Append" value="true" /> <layout class="org.apache.log4j.PatternLayout"> <param name="ConversionPattern" value="[%t] %d{yyyy-MM-dd HH:mm:ss.SSS} %-5p %c{1} - %m%n" /> </layout> </appender> <root> <priority value ="info" /> <appender-ref ref="gemFile" /> </root> </log4j:configuration> ${catalina.home} resolves to C:\Apps\Apache\Tomcat7. When I deploy my WAR file, I get log4j:ERROR setFile(null,true) call failed. java.io.FileNotFoundException: file:C:\Apps\Apache\Tomcat7\logs\GEM.log (The filename, directory name, or volume label syntax is incorrect) stack trace... It is true that the log file does not exist (the folder structure does), but if I set the file directly to that path (<param name="File" value="C:/Apps/Apache/Tomcat7/logs/GEM.log" />) it works fine. Why can't I use the replacement string?

    Read the article

  • Use netcat as a proxy to log traffic

    - by deephacks
    I want to use netcat as a proxy to log http requests and responses to files, then tail these to inspect traffic. Think wireshark. Tried the following where 'fifo' is a named pipe, 'in' and 'out' are files, netcat proxy on port 8080, server on port 8081. while true; do cat fifo | nc -l -p 8080 | tee -a in | nc localhost 8081 | tee -a out 1fifo; done Problems: Netcat stop responing after first request (while loop ignored?). Netcat fails with msg localhost [127.0.0.1] 8081 (tproxy) : Connection refused if server unavailable on 8081. Question: Is it possible to "lazily" connect to 8081 when request is made? I.e. I do not want to have 8081 running when netcat is started.

    Read the article

  • log4net creates log file but does not write to it (windows service in C#)

    - by user1825172
    I am trying to use basic logging for a windows service. I added the reference to log4net I added the following in AssemblyInfo.cs: [assembly: log4net.Config.XmlConfigurator(Watch = true)] I added the following to my App.config: <?xml version="1.0" encoding="utf-8" ?> <configuration> <configSections> <section name="log4net" type="log4net.Config.Log4NetConfigurationSectionHandler,log4net, Version=1.2.10.0, Culture=neutral, PublicKeyToken=1b44e1d426115821" requirePermission="false" /> </configSections> <!-- Log4net Logging Setup --> <log4net> <appender name="RollingFileAppender" type="log4net.Appender.RollingFileAppender,log4net"> <file value="c:\\CGSD\\log\\logfile.txt" /> <appendToFile value="true" /> <lockingModel type="log4net.Appender.FileAppender+MinimalLock" /> <layout type="log4net.Layout.PatternLayout"> <conversionPattern value="%date [%thread] %level %logger - %message%newline" /> </layout> <filter type="log4net.Filter.LevelRangeFilter"> <levelMin value="INFO" /> <levelMax value="FATAL" /> </filter> </appender> <root> <level value="ALL"/> <appender-ref ref="RollingFileAppender"/> </root> </log4net> </configuration> I have the following code in my service: log4net.Config.XmlConfigurator.Configure(); log4net.ILog log = log4net.LogManager.GetLogger(typeof(Program)); log.Debug("test"); the file c:\CGSD\log\logfile.txt is created but nothing is ever written to it. i've been thru the forums all day trying to trac this one down, but if i overlooked an already posted solution i apologize. thx!

    Read the article

  • Pidgin: a me-specific log file?

    - by Wolfpack'08
    I'm trying to make a compilation file of all the text sent by me and all the responses that included my name.... What's the best way about scripting this? Is there a plugin like that for pidgin? I'd also like to be able to click on lines to add. For example, someone gives a really useful, easy-to-follow guide to another user, and I think: "I'm going to be doing that tomorrow." Rather than marking and pasting every line of it and skipping through all the other convo, I would like to be able to just click 'send to my personal log,' or something. It seems quite useful, right? And I think that a lot of people would use it. So, please try to provide a comprehensive answer, if possible, or one that I can follow through on.

    Read the article

  • Apache Error Log - "Web Path" instead of Filesystem Path

    - by Craconia
    Hello everyone, I'm running Apache on Linux and I'm using OpenSSH to provide SFTP access to some customers so they can upload their pages and also look at their respective site logs (access & error). I'm using the new feature in OpenSSH to chroot their SFTP access and so far so good. My problem is that on the error_log, every reference for "File not found..." is given using the OS filesystem path as opposed to the "Web" path. I'd rather have the web path on the error log in order not to reveal the OS path. Since I'm already chrooting the users, I don't want to reveal WHERE on the OS their files are actually located... Is it possible to change this behaviour via any directive? I tried looking for it but couldn't find anything :( Thanks, Craconia

    Read the article

  • Delay at Windows Log On Screen

    - by Ryan Elkins
    I'm getting a delay when entering my password at the log on screen in Windows 7. What I mean is that I can select the text box but pressing keys does nothing. I had this problem initially on this computer with Vista installed - when I reformatted and installed Windows 7 the problem went away - but over the weeks it is slowly coming back (the delay seems to be getting longer). Any clues as to why this is happening and how I can fix it? It currently takes about 4-8 seconds before it will accept keyboard input.

    Read the article

  • Perf4J Graph Output from Log File

    - by manyxcxi
    I currently have a long running process that I am trying to analyze with Perf4J. I currently have it writing results in CSV format to its own log file using the AsyncCoalescingStatisticsAppender and a StatisticsCsvLayout on the file appender. My question is; when I try and use the --graph option from the command line (using the perf4j jar) it isn't populating the data points- it isn't populating anything. Are my appenders set incorrectly? The log file contains hundreds (sometimes thousands) of data points of about 10 different tag names. <appender name="perfAppender" class="org.apache.log4j.FileAppender"> <param name="File" value="perfStats.log"/> <layout class="org.perf4j.log4j.StatisticsCsvLayout"> </layout> </appender> <appender name="CoalescingStatistics" class="org.perf4j.log4j.AsyncCoalescingStatisticsAppender"> <!-- The TimeSlice option is used to determine the time window for which all received StopWatch logs are aggregated to create a single GroupedTimingStatistics log. Here we set it to 10 seconds, overriding the default of 30000 ms --> <param name="TimeSlice" value="10000"/> <appender-ref ref="ConsoleAppender"/> <appender-ref ref="CompositeRollingFileAppender"/> <appender-ref ref="perfAppender"/> </appender>

    Read the article

  • Error # 1045 - Cannot Log in to MySQL server -> phpmyadmin

    - by SilverLight
    We have installed PHPMyAdmin on a windows machine running IIS 7.0. We are able to connect to MySQL using command-line, But we are not able to connect using PHPMyAdmin. The error displayed is: Error #1045 Cannot log in to the MySQL server. Can somebody please help? PHP Version 5.4.0 mysqlnd 5.0.10 - 20111026 - $Revision: 323634 $ phpMyAdmin-3.5.4-rc1-all-languages.7z EDIT : I followed the link below with no success, mean i changed that password but phpmyadmin still has that error... C.5.4.1.1. Resetting the Root Password: Windows Systems Thanks.

    Read the article

  • Win 7 laptop taking a long time to log on when off the domain

    - by pigeon
    I've clicked through on a few forums but unable to find a suggestion for this issue. I have a user with a HP laptop which will boot perfectly fine whilst connected to the network but at home it will take 4 minutes to log on. Some suggestions I have had is that its looking for the DNS server of the domain. Another was that it was suggested it was a roaming profile (it isn't). So hoping someone has encountered Windows 7 acting like this and has some suggestions. The other users on this domain don't have the issue and its a mix of XP and Win 7 client PCs.

    Read the article

  • How do I combine multiple changelog / commit entries in mercurial

    - by Kimvais
    I currently have a repository, where development work has been committed directly into the 'default' branch. I'd like to find a way to combine multiple small changesets into a single changeset, for merging the changes to child repositories. E.g in the default/main repo there are changesets something like: 12839: Fooed the bar 12838: Fixed blonking 12837: Trumped slamdunks ... 323: Started development for thingamajiks and I would like this to look like (in the child repos): 323: Added thingamajik functionality

    Read the article

  • syslog-ng fails to log on lxc host

    - by christian
    we are running CentOS 6 servers with multiple lxc-containers. For system logging we are using syslog-ng. After a while the syslog-ng daemon stops logging messages, but the daemon keeps running. This happens on the host and inside the containers (where another syslog-ng is running) as well. We could not find any patterns for the failure yet but we assume that it has something to do with lxc, because we don't have these problems on other hosts. We have the suspicion that these problems occur when more than on lxc-container is running and that only "new" processes can not log. We are running the following software versions: CentOS-Linux 6.4/6.5 lxc-0.7.5 syslog-ng-3.2.5 Do you have any ideas? Best regards trademesh

    Read the article

  • Java: library that does nice formatted log outputs

    - by WizardOfOdds
    I cannot find back a library that allowed to format log output statements in a much nicer way than what is usually seen. One of the feature I remember is that it could 'offset' the log message depending on the 'nestedness' of where the log statement was occuring. That is, instead of this: DEBUG | DefaultBeanDefinitionDocumentReader.java| 86 | Loading bean definitions DEBUG | AbstractAutowireCapableBeanFactory.java| 411 | Finished creating instance of bean 'MS-SQL' DEBUG | DefaultSingletonBeanRegistry.java| 213 | Creating shared instance of singleton bean 'MySQL' DEBUG | AutowireCapableBeanFactory.java| 383 | Creating instance of bean 'MySQL' DEBUG | AutowireCapableBeanFactory.java| 459 | Eagerly caching bean 'MySQL' to allow for resolving potential circular references DEBUG | AutowireCapableBeanFactory.java| 789 | Another debug message It would shows something like this: DEBUG | DefaultBeanDefinitionDocumentReader.java| 86 | Loading bean definitions DEBUG | AbstractAutowireCapableBeanFactory.java | 411 | Finished creating instance of bean 'MS-SQL' DEBUG | DefaultSingletonBeanRegistry.java | 213 | Creating shared instance of singleton bean 'MySQL' DEBUG | AutowireCapableBeanFactory.java | 383 | Creating instance of bean 'MySQL' DEBUG | AutowireCapableBeanFactory.java | 459 | |__ Eagerly caching bean 'MySQL' to allow for resolving potential circular references DEBUG | AutowireCapableBeanFactory.java | 789 | |__ Another debug message This is an example I just made up (VeryLongCamelCaseClassNamesNotMine). But I remember seeing such cleanly formatted log output and they were really much nicer than anything I had seen before and, in addition to being just plain nicer, they were also easier to read for they reproduced some of the logical organization of the code. Yet I cannot find anymore what that library was. I'm pretty sure it was fully compatible with log4j or sl4j.

    Read the article

  • How to set permissions or alter a git commit process when using local repositories

    - by Tony
    I have a server that contains a central git repository and one of my co-worker's development environment. My co-worker's repository's origin is the central git repository and he pushes there when he has some code to share. Likewise, I develop locally and push to the central git repository when I have some code to share, so my repository's origin is also the central git repository. The issue is that I have the central git repository under a "git" user's home directory. So when I push I am actually SSH'ing into the the server as the "git" user. To be even more clear, my config has these lines: $ more .git/config [remote "origin"] fetch = +refs/heads/*:refs/remotes/origin/* url = [email protected]:fsg [branch "master"] remote = origin merge = refs/heads/master When I push, git handles this SSH + push seamlessly with I am guessing some sort of git shell. The issue is that when my coworker pushes, he is logged in as himself for a user and gets a bunch of crazy permission errors. Is there a typical way to solve this problem without opening up git's directories to a group? I think this will be problematic when I push and therefore overwrite the the repository and those permissions are reset. Thanks!

    Read the article

  • Download a website that requires log-in with HTTtrack Copier

    - by H.Moss
    Hi guys! I have been researching of how to download content of a site that requires username and password. This is actually harder than I thought it would be. I tried to use both HTTtrack Copier and followed the instruction below, but it's not working! Q: I can not access several pages (access forbidden, or redirect to another location), but I can with my browser, what's going on? A: You may need cookies! Cookies are specific data (for example, your username or password) that are sent to your browser once you have logged in certain sites so that you only have to log-in once. For example, after having entered your username in a website, you can view pages and articles, and the next time you will go to this site, you will not have to re-enter your username/password. To "merge" your personnal cookies to an HTTrack project, just copy the cookies.txt file from your Netscape folder (or the cookies located into the Temporary Internet Files folder for IE) into your project folder (or even the HTTrack folder)

    Read the article

  • Sending Subversion Change Log Info Via Hudson

    - by GrumpyCanuck
    I'm trying to integrate Hudson into our development process, and everything is going smooth except for one thing. I had been using Phing to do deployments, and one of the things that was being triggered was an email to our tech support email address containing a list of all the commit messages between the last time code was deployed and the present SVN revision. I was doing something like this: read in a file from the root directory of the currently-deployed application that contains the SVN revision when the app was deployed place that value in a Phing variable insert that value into a command to send the SVN commit messages via email create a file in the root directory of the newly-deployed application that contains the current SVN revision I'd like to be able to add that information to the email that gets sent out by Hudson when a successful build goes out. Any pointers on how to accomplish this task in Hudson would be greatly appreciated.

    Read the article

  • how to log trigger intiated CRUD queries on postgresql 8.4

    - by user47650
    in the postgresql config file there is an option log_statement = 'mod' which causes CRUD statements to be logged. However this does not include the CRUD statements that are called from trigger functions, which means that following the log file is not useful to determine what changes are being made to the database data, if the 3rd part application has been making lots of changes using triggers. Is there some other option I can use to include trigger CRUD? alternatively, can I inspect the pg_xlog in real time using some tool? (xlogdump and xlogviewer do not work with version 8.4, i have tried)

    Read the article

< Previous Page | 42 43 44 45 46 47 48 49 50 51 52 53  | Next Page >