Search Results

Search found 11703 results on 469 pages for 'dev to production'.

Page 293/469 | < Previous Page | 289 290 291 292 293 294 295 296 297 298 299 300  | Next Page >

  • How many javascript link I can have in Facebook?

    - by Murvinlai
    Is there a limit of how many js file I can include () in facebook? I include 5 files. no problem.. the 6th one not loaded. Then I have to put the code in the 6th one into the 5th file. then works. so, is it 5 files in max? BY THE WAY, I'm developing the apps now, not in production. so it is not in the stage of compressing JS / minizing it. :) so it is kinda annoying to got missing js files or files not loaded..etc so, what is the limit from FB? what is the file size limit? I know that the JSON request call back data limit is 5000.. but not sure about the js include.

    Read the article

  • Microsoft.Biztalk.explorerom.dll reference in asp.net application resulting system.nullreferenceexce

    - by sheetal.oza
    Hi, I have a asp.net application in order to start/stop applications and ports of Biztalk server 2006 r2. I have used "Microsoft.Biztalk.explorerom.dll (C:/Program Files/Biztalk Server 2006/Developer tool) " to achieve this. This is working fine on development machine since biz talk server is installed on local machine. But in the production environment (asp.net web server ,windows 2003 and iis 6.0)...this give System.nullreferenceexception (object reference not set..) at BtsCatalogExplorer explorer = (BtsCatalogExplorer)myGroup.CreateInstance(typeof(BtsCatalogExplorer)) my biztalk server and sql server are on two different box. In my setup (asp.net web application)..adding Microsoft.Biztalk.explorerom.dll and Microsoft.Biztalk.Applicationdeployment.engine.dll to GAC. But still no luck. Do i need to install biz talk server on my local machine even though I am connecting to different biz talk server?? And help is appreciated...

    Read the article

  • Reference app relative virtual paths in .css file

    - by bravo9
    Assume I have an "images" folder directory under the root of my application. How can I, from within a .css file, reference an image in this directory using an ASP.NET app relative path. Example: When in development, the path of ~/Images/Test.gif might resolve to /MyApp/Images/Test.gif while, in production, it might resolve to /Images/Test.gif (depending on the virtual directory for the application). I, obviously, want to avoid having to modify the .css file between environments. I know you can use Page.ResolveClientUrl to inject a url into a control's Style collection dynamically at render time. I would like to avoid doing this.

    Read the article

  • Is there a way to read what NSLog is writing from within the app?

    - by Jared P
    I have already tried stderr = stdout = fopen("/Users/Jared/Desktop/Untitled.txt", "w"); just as a test (obvi not production code) and this works fine for printf and some errors, but fails to redirect NSLog. I would like to be able to read what is being sent to NSLog as a string, and preferably prevent it from actually going into the system logs, although I don't particularly care if it has to. I need it to redirect logging from both my own code and apple's frameworks, which is why I can't just write a function that would append logs to a string. Thanks

    Read the article

  • inject a mockups to a bean that has @autowired annotations

    - by santiagozky
    I have a bean that has a couple of beans injected with the autowire annotation (no qualifier). Now, for testing reasons I want to inject some mocks to the bean instead of the ones being autowired (some DAOs). Is there any way I can change which bean is being injected without modifying my bean? I don't like the idea of adding annotations my code just to test it and then remove then for production. I am using spring 2.5. The bean look like this: @Transactional @Service("validaBusinesService") public class ValidaBusinesServiceImpl implements ValidaBusinesService { @Autowired OperationDAO operationDAO; @Autowired BinDAO binDAO; @Autowired CardDAO cardDAO; @Autowired UserDAO userDAO; ... ... }

    Read the article

  • DbDataReader with DbTransactions

    - by Gustavo Paulillo
    Its the wrong way or lack of performance, using DbDataReader combinated with DbTransactions? An example of code: public DbDataReader ExecuteReader() { try { if (this._baseConnection.State == ConnectionState.Closed) this._baseConnection.Open(); if (this._baseCommand.Transaction != null) return this._baseCommand.ExecuteReader(); return this._baseCommand.ExecuteReader(CommandBehavior.CloseConnection); } catch (Exception excp) { if (this._baseCommand.Transaction != null) this._baseCommand.Transaction.Rollback(); this._baseCommand.CommandText = string.Empty; this._baseConnection.Close(); throw new Exception(excp.Message); } } Some methods call this operation. Sometimes openning a DbTransaction. Its using DbConnection and DbCommand. The real problem, is in production enviroment (like 5,000 access/day) the ADO operations start throwing exceptions

    Read the article

  • MS SQL Server: how to optimize "like" queries?

    - by duke84
    I have a query that searches for clients using "like" with wildcard. For example: SELECT TOP (10) [t0].[CLIENTNUMBER], [t0].[FIRSTNAME], [t0].[LASTNAME], [t0].[MI], [t0].[MDOCNUMBER] FROM [dbo].[CLIENT] AS [t0] WHERE (LTRIM(RTRIM([t0].[DOCREVNO])) = '0') AND ([t0].[FIRSTNAME] LIKE '%John%') AND ([t0].[LASTNAME] LIKE '%Smith%') AND ([t0].[SSN] LIKE '%123%') AND ([t0].[CLIENTNUMBER] LIKE '%123%') AND ([t0].[MDOCNUMBER] LIKE '%123%') AND ([t0].[CLIENTINDICATOR] = 'ON') It can also use less parameters in "where" clause, for example: SELECT TOP (10) [t0].[CLIENTNUMBER], [t0].[FIRSTNAME], [t0].[LASTNAME], [t0].[MI], [t0].[MDOCNUMBER] FROM [dbo].[CLIENT] AS [t0] WHERE (LTRIM(RTRIM([t0].[DOCREVNO])) = '0') AND ([t0].[FIRSTNAME] LIKE '%John%') AND ([t0].[CLIENTINDICATOR] = 'ON') Can anybody tell what is the best way to optimize performance of such query? Maybe I need to create an index? This table can have up to 1000K records in production.

    Read the article

  • JSON webservice response compression in IIS 7

    - by denisioru
    Hello! I have trouble with JSON response compression. I look to response headers after uploading website to production server (Windows 2008, IIS 7) and found uncompressed response. Turning on "Enabled static compression" and ""Enable dynamic compression" in IIS control panel does not effect. ASPX pages was responsed gzipped, but webservice response uncompressed. I looked to google, but no answer found about this trouble. Also, I try this http://stackoverflow.com/questions/2405595/json-ihttpmodule-compression way (and adding to web.config this module) - but this source is excellent working at development machine with ASP.NET development server (and have seven times response size reduced) and totally ignored at IIS7. How I can apply gzip compression to json responses from my webservice? Thanks. PS .NET 3.5

    Read the article

  • Changing the Column order/adding Newcolumn for existing Table in SQLServer2008

    - by rmdussa
    Hi I have situation where I need to change the order of the columns/adding new columns for existing Table in SQLServer2008, It is not allowing me to do without drop and recreate. But That is in production system and having data in that table. I can take backup of the data, and Drop the existing table and change the order/add new columns and recreate it, insert the backup data into new Table. Is there any best way to do this without dropping and recreating.I think SQLServer 2005 will allow this process without dropping and recreating while changing to existing table structure. Thanks

    Read the article

  • Linked Server related

    - by rmdussa
    I have two instances of SQL Server: Server1 (SQL Server 2008) Server2 (SQL Server 2005) I am executing a stored procedure from Server1 which references tables on Server2. It is working fine in my test environment: Server1 runs Vista SP2, SQL Server 2008; Server2 runs Windows XP SP2, SQL Server 2005. However, it is not working in the production environment: Server1 runs Vista SP1, SQL Server 2008; Server2 runs Windows XP SP2, SQL Server 2005. The error message I receive is: OLE DB provider "SQLNCLI10" for linked server "Server2" returned message "No transaction is active.". Msg 7391, Level 16, State 2, Line 21 The operation could not be performed because OLE DB provider "SQLNCLI10" for linked server "Server2" was unable to begin a distributed transaction.

    Read the article

  • PDO and SQL IN statements

    - by Sai
    Im using a sequel for search like this using PDOs $states = "'SC','SD'"; $sql = "select * from mytable where states in (:states)"; $params = array(':states'=>$states); and I use my function $result = $this->selectArrayAssoc($sql, $params); where my selectArrayAssoc function as following public function selectArrayAssoc($sql, $params = array()){ try{ $sth = $this->db->prepare($sql); $sth->execute($params); $result = $sth->setFetchMode(PDO::FETCH_ASSOC); return $sth->fetchAll(); }catch(PDOException $e){ print $e->getMessage(); //Log this to a file later when in production exit; } } it does not take the quoted variables, I think it is suppressing, in such cases how to deal with this.

    Read the article

  • SQL Server: how to optimize "like" queries?

    - by duke84
    I have a query that searches for clients using "like" with wildcard. For example: SELECT TOP (10) [t0].[CLIENTNUMBER], [t0].[FIRSTNAME], [t0].[LASTNAME], [t0].[MI], [t0].[MDOCNUMBER] FROM [dbo].[CLIENT] AS [t0] WHERE (LTRIM(RTRIM([t0].[DOCREVNO])) = '0') AND ([t0].[FIRSTNAME] LIKE '%John%') AND ([t0].[LASTNAME] LIKE '%Smith%') AND ([t0].[SSN] LIKE '%123%') AND ([t0].[CLIENTNUMBER] LIKE '%123%') AND ([t0].[MDOCNUMBER] LIKE '%123%') AND ([t0].[CLIENTINDICATOR] = 'ON') It can also use less parameters in "where" clause, for example: SELECT TOP (10) [t0].[CLIENTNUMBER], [t0].[FIRSTNAME], [t0].[LASTNAME], [t0].[MI], [t0].[MDOCNUMBER] FROM [dbo].[CLIENT] AS [t0] WHERE (LTRIM(RTRIM([t0].[DOCREVNO])) = '0') AND ([t0].[FIRSTNAME] LIKE '%John%') AND ([t0].[CLIENTINDICATOR] = 'ON') Can anybody tell what is the best way to optimize performance of such query? Maybe I need to create an index? This table can have up to 1000K records in production.

    Read the article

  • Export Plone users to LDAP?

    - by Kevin
    Hi, I've been tasked with setting up a subversion server next to a Plone instance. The situation is that the Plone instance is already in a "production" state because there are outside users that can access it any moment. I believe setting up a LDAP and binding both, the Plone and subversion instance, to it for authentication is the best solution. So, my question is "how do I export already existing users and respective credentials from a Plone instance to an LDAP install for a seamless transition?" Thanks for any suggestions in advance!

    Read the article

  • How to log into a salesforce.com sandbox?

    - by wesgarrison
    I took over a Rails app and am trying to get the Salesforce.com API credentials set up for my user account. I'm set as a system admin with "Developer Mode" on (though I have no clue what that does, I just saw it set on the previous account.) We have a sandbox. I click the login link on it, enter my sandbox username ([email protected]), enter my password, and get "Login attempt has failed". I know my password is okay since I've logged into the production site several times. Using different domains (test.salesforce.com, etc) doesn't help either.

    Read the article

  • ELMAH with IIS6.0 and SQL Server 2005

    - by RajeshT
    On production I have ELMAH configured so that the errors are logged into a SQL Server 2005 database created specially for ELMAH. The sites run on IIS6.0 web server. ELMAH is able to send emails for the errors, but nothing is getting logged into the SQL Server database. The connection is a trusted connection and I have <section name="errorLog" requirePermission="false" type="Elmah.ErrorLogSectionHandler, Elmah"/> in section group <errorLog type="Elmah.SqlErrorLog, Elmah" connectionStringName="Elmah.Sql" applicationName="VBCPortal"/> under elmah <add name="Elmah.Sql" connectionString="Data Source=(local);Initial Catalog=ELMAH;Trusted_Connection=True" /> in connectionStrings in my web.config file. Any idea why the errors are not getting logged into the database?

    Read the article

  • Migrating VseWss project to WspBuilder

    - by Khurram Aziz
    I have a VseWss project that I want to migrate to WspBuilder. The project has many features (Content Types, Fields, Lists, Event Listeners, Workflows, ASPX files, their code behind etc) and the project references couple of other assemblies as well. And it does the SafeControls entries and deploys the assemblies into GAC etc Important aspect is; the WSP that gets created from the VseWss is deployed to the production server and while migrating I want to preserve all of these things so that when I create the WSP from WspBuilder things are transparent for the Sharepoint Administrators. Please guide me how to proceed...any online tutorial or related stuff that talks about migrating a real world VseWss project (with 10-20 different types of features)

    Read the article

  • Using WebParts results in blank page

    - by dale
    Hi, I'm developing an ASP.NET (C#) application using EPiServer CMS 5. On the startpage we have four WebPart-zones and everything works great running it through Visual Studio. When publishing it to the production-server the startpage shows as a blank page, no error what so ever... When we remove these lines of code it works, so it has something to do with the WebParts. <WebParts:ExtendedWebPartManager ID="WebPartsManager" runat="server"></WebParts:ExtendedWebPartManager> <WebParts:ExtendedWebPartZone WebPartProperty="WebPartZone1" runat="server" ID="WebPartZone1" PartChromeType="None" AllowLayoutChange="false" Padding="0" PartChromePadding="0" Width="100%"></WebParts:ExtendedWebPartZone> <WebParts:ExtendedWebPartZone WebPartProperty="WebPartZone2" runat="server" ID="WebPartZone2" PartChromeType="None" AllowLayoutChange="false" Padding="0" PartChromePadding="0" Width="100%"></WebParts:ExtendedWebPartZone> <WebParts:ExtendedWebPartZone WebPartProperty="WebPartZone3" runat="server" ID="WebPartZone3" PartChromeType="None" AllowLayoutChange="false" Padding="0" PartChromePadding="0" Width="100%"></WebParts:ExtendedWebPartZone> <WebParts:ExtendedWebPartZone WebPartProperty="WebPartZone4" runat="server" ID="WebPartZone4" PartChromeType="None" AllowLayoutChange="false" Padding="0" PartChromePadding="0" Width="100%"></WebParts:ExtendedWebPartZone>

    Read the article

  • Expression Engine Template Exporting

    - by Lark
    I have a project that was developed into a Expression Engine template. I used the Core version of EE to develop this template on my local machine. Now I need to export this template/weblog so it can be loaded onto the production server so the client can begin using EE to update their site. The issue I am running into is how to export the templates plus the weblogs that go along with the template. This is my first time using EE so if you know how to accomplish this please explain it in detail. Thank you in advance.

    Read the article

  • Attachment_fu error

    - by cswebgrl
    Hello, I am getting an error while trying to upload images on an Ubuntu machine that's running Rails 2.3.4, Ruby 1.8.6 using attachment_fu with image science. FreeImage exception for type ???: IPTC: Invalid key 'Tag 0x025C' The error seems to point to this line in the image_science_processor in the attachment_fu plugin: def with_image(file, &block) ::ImageScience.with_image file, &block end My initial thoughts are that it has something to do with meta tags and the images and maybe free image. I don't actually see this error on my dev machine - Mac Snow Leopard, Rails 2.3.5, Ruby 1.8.7. Before I start messing versions on the production boxes, has anyone else encountered this issue and have an idea to fix it? THANKS!!!!

    Read the article

  • Rails: blank page - no errors or stack trace

    - by Nathan Long
    I've been trying to fix a bug in the Rails app I'm developing, and I keep getting a blank screen with no errors. I haven't found anything helpful in development.log, either (though it does show queries being run and such). Finally, I started to wonder if it's somehow set not to show errors anymore. I tried commenting out a necessary route, and sure enough, I got a blank page instead of the error and stack trace I expected. What might cause this? (I wondered if maybe I'm accidentally running production mode and errors aren't supposed to show then, but development.log is being appended, and if I open script/console and echo ENV['RAILS_ENV'], it says development.)

    Read the article

  • How to set AeDebug to get a minidump with the name of the process ?

    - by JC Martin
    I have to perform some post mortem debugging on a C++ project. Known way to perform is to set the cdb debugger as a minidump generator and to process the dumps collects afterwards. I read nearly the whole web and I didn't find a solution to produce a minidump with the name of the process that has crashed Is there a way to set AeDebug\Debugger registry variable in such a manner that cdb generates a dump file with the name of the process ? When I encapsulate the call to cdb.exe in a batch file, it starts well but stays blocked on the symbol searching. I must perform a Ctrl+C in order to stop the batch, then the minidump, with the correct process name, is created... but of course I can't set up such a thing in an unattended production environment... Has anybody done that before ?

    Read the article

  • Getting HTTP 406 when trying to test facebooker application with cucumber

    - by Waseem
    I am trying to test facebook api calls with cucumber. Here is the code. # app/controller/facebook_users_controller.rb class FacebookUsersController < ApplicationController def create fb_user = facebook_session.user user = User.new(:facebook_uid => fb_user.uid, :facebook_session_key => facebook_session.session_key respond_to do |format| if user.save format.json { render :json => { :status => 'ok' }.to_json } end end end end # features/steps/facebook_connect_step.rb Given /^I am a facebook connected user$/ do mock_session = Facebooker::MockSession.create post('/facebook_user.json') puts response.code end When I run the cucumber step for above step definition, I get a response code of 406 instead of 200. This happens in the cucumber test environment only and not in the browser(development/production).

    Read the article

  • PHP's apache_setenv function causes 500 Internal Server Error

    - by guitar-
    apache_setenv ( 'no-gzip', 1 ) I'm trying to disable gzip for a certain page's output, but only that page. This works fine on testing servers, but not the production server, which is running the same thing (CentOS and Apache), works on Ubuntu though. Anyway, do you know why? Or is there some other alternative? I was thinking of using ob_start () to put all output in a buffer, and then unzip it myself with a PHP function then call ob_end_flush ()... or would it not be gzipped until right before Apache sends it to the client? Thanks for any help.

    Read the article

  • ASP.NET MVC TDD with LINQ and SQL database

    - by dean nolan
    I am trying to start a new MVC project with tests and I thought the best way to go would have 2 databases. 1 for testing against and 1 for when I run the app and use it (also test really as it's not production yet). For the test database I was thinking of putting create table scripts and fill data scripts within the test setup method and then deleting all this in the tear down method. I am going to be using Linq to SQL though and I don't think that will allow me to do this? Will I have to just go the ADO route if I want to do it this way? Or should I just use a mock object and store data as an array or something?. Any tips on best practices? How did Jeff go about doing this for StackOveflow?

    Read the article

  • Good way to capture/replay sessions from Apache Log?

    - by Mark Harrison
    For performance testing, I would like to capture some traffic from a production server and use that as a basis to replay the request to a test server in order to simulate a realistic load in our development environment. These are all stateless queries, so no issues regarding cookies, sessions, etc. The Apache log timestamps everything down to a 1 second resolution, but that's not fine enough granularity for our peak times. What's the best way to capture more fine-grained timestamps for replay? And is there some ab-like load generating program that can use this data to replicate load?

    Read the article

< Previous Page | 289 290 291 292 293 294 295 296 297 298 299 300  | Next Page >