Search Results

Search found 5144 results on 206 pages for 'unsupported packages'.

Page 80/206 | < Previous Page | 76 77 78 79 80 81 82 83 84 85 86 87  | Next Page >

  • Python Eggs on Google App Engine

    - by ott
    Usually I would use virtualenv and pip for deployment of web applications. With Google App Engine this doesn't work, because all import statement are relative to directory of the application. The most common approach I saw was to simply copy the packages from site-packages to the directory of the application. This involves manual work and is error-prone. Another approach was to changes install_lib and install_scripts in ~/.pydisutils.cfg, but this doesn't allow me to use pip in my home directory simultaneously. Do you have any suggestions for this?

    Read the article

  • What does the "build-essential" Terminal command mean & do in Linux based operating systems like Ubu

    - by Adam Siddhi
    Hi. I am researching how to install Ruby 1.9.1 in Xubuntu 10.04 and I came across the command build-essential multiple times. Sometimes it is followed by packages and sometimes it is both preceded and post-ceded by packages. The 2 examples I am looking at are: sudo apt-get install build-essential zlib1g zlib1g-dev zlibc libruby1.9 libxml2 libxml2-dev libxslt-dev and sudo apt-get install ruby irb ri rdoc ruby1.8-dev libzlib-ruby libyaml-ruby libreadline-ruby libncurses-ruby libcurses-ruby libruby libruby-extras libfcgi-ruby1.8 build-essential libopenssl-ruby libdbm-ruby libdbi-ruby libdbd-sqlite3-ruby sqlite3 libsqlite3-dev libsqlite3-ruby libxml-ruby libxml2-dev Thanks :adam

    Read the article

  • How to import a module from a directory?

    - by Roman
    On my system I have two versions of Python (to call them I type python and python2 in the command line). When I use the first version of Python, I cannot import sklearn module but I can do it in the second version of Python. I would like to use the first version of python (because other modules are available there) and, at the same time, I would like to be able to import sklearn from this version of Python. My solution was to use: import sys sys.path.append('location_of_the_sklearn_module') To find the location of the sklearn module I started a python session (using the second version of python, in which sklearn works). The I type: import sklearn sklearn.__file__ As a result I got: /home/name/my_name/numpy/local/lib/python2.7/site-packages/sklearn/__init__.pyc In the session of the first version of Python I tried: import sys sys.path.append('/home/name/my_name/numpy/local/lib/python2.7/site-packages/sklearn') import sklearn Unfortunately it did not work. As a result I got: ImportError: No module named sklearn Does anybody know what I am doing wrong and if it is possible to reach the goal in the way I try?

    Read the article

  • What database table structure should I use for versions, codebases, deployables?

    - by Zac Thompson
    I'm having doubts about my table structure, and I wonder if there is a better approach. I've got a little database for version control repositories (e.g. SVN), the packages (e.g. Linux RPMs) built therefrom, and the versions (e.g. 1.2.3-4) thereof. A given repository might produce no packages, or several, but if there are more than one for a given repository then a particular version for that repository will indicate a single "tag" of the codebase. A particular version "string" might be used to tag a version of the source code in more than one repository, but there may be no relationship between "1.0" for two different repos. So if packages P and Q both come from repo R, then P 1.0 and Q 1.0 are both built from the 1.0 tag of repo R. But if package X comes from repo Y, then X 1.0 has no relationship to P 1.0. In my (simplified) model, I have the following tables (the x_id columns are auto-incrementing surrogate keys; you can pretend I'm using a different primary key if you wish, it's not really important): repository - repository_id - repository_name (unique) ... version - version_id - version_string (unique for a particular repository) - repository_id ... package - package_id - package_name (unique) - repository_id ... This makes it easy for me to see, for example, what are valid versions of a given package: I can join with the version table using the repository_id. However, suppose I would like to add some information to this database, e.g., to indicate which package versions have been approved for release. I certainly need a new table: package_version - version_id - package_id - package_version_released ... Again, the nature of the keys that I use are not really important to my problem, and you can imagine that the data column is "promotion_level" or something if that helps. My doubts arise when I realize that there's really a very close relationship between the version_id and the package_id in my new table ... they must share the same repository_id. Only a small subset of package/version combinations are valid. So I should have some kind of constraint on those columns, enforcing that ... ... I don't know, it just feels off, somehow. Like I'm including somehow more information than I really need? I don't know how to explain my hesitance here. I can't figure out which (if any) normal form I'm violating, but I also can't find an example of a schema with this sort of structure ... not being a DBA by profession I'm not sure where to look. So I'm asking: am I just being overly sensitive?

    Read the article

  • Run SSIS Package from T-SQL

    - by Dr. Zim
    I noticed you can use the following stored procedures (in order) to schedule a SSIS package: msdb.dbo.sp_add_category @class=N'JOB', @type=N'LOCAL', @name=N'[Uncategorized (Local)]' msdb.dbo.sp_add_job ... msdb.dbo.sp_add_jobstep ... msdb.dbo.sp_update_job ... msdb.dbo.sp_add_jobschedule ... msdb.dbo.sp_add_jobserver ... (You can see an example by right clicking a scheduled job and selecting "Script Job as- Create To".) AND you can use sp_start_job to execute the job immediately, effectively running SSIS packages on demand. Question: does anyone know of any msdb.dbo.[...] stored procedures that simply allow you to run SSIS packages on the fly without using sp_cmdshell directly, or some easier approach?

    Read the article

  • associating a filetype with a batch script, and getting parameters passed to file of that type.

    - by Carson Myers
    Sorry for the cryptic title. I have associated python scripts with a batch file that looks like this: python %* I did this because on my machine, python is installed at C:\python26 and I prefer not to reinstall it (for some reason, it won't let me add a file association to the python interpreter. I can copy the executable to Program Files and it works -- but nothing out of Program Files seems to work). Anyways, I can do this, so far: C:\py django-admin C:\py python "C:\python26\Lib\site-packages\django\bin\django-admin.py" Type 'django-admin.py help' for usage. C:\py django-admin startproject myProj C:\py python "C:\python26\Lib\site-packages\django\bin\django-admin.py" Type 'django-admin.py help' for usage. but the additional parameters don't get passed along to the batch script. This is getting very annoying, all I want to do is run python scripts :) How can I grab the rest of the parameters in this situation?

    Read the article

  • "import as" leads to unresolved import error, "from .. import" does not

    - by Markus R.
    I'm trying to write some plugins for the irc bot supybot with eclipse/pydev. Pydev gives me errors about unresolved imports on supybot-modules/packages (e. g. import supybot.utils as utils), but works ok on e. g. "from supybot.commands import *". So I guess I set up dydev correctly, as it finds the wanted modules. The problem must be in pydev/eclipse, as the bot works correct and in eric5 I get also no errors about that. Removing the interpreter and setting it up didn't help. Any other ideas on how to fix this? System: Arch Linux, Eclipse Juno, PyDev 2.7.1, wanted (and set up) python interpreter is 2.7, supybot is installed in site-packages for Python 2.7. Edit: Just noticed: PyDev doesn't mark the "from ... import *" as error, but if I use functions imported from there I get an error on that function.

    Read the article

  • Where do files included in MANIFEST.in end up?

    - by Brian Hicks
    I'm not sure if I can't find this or if my google-fu is just lacking at the moment: I've got some HTML template files included in a package, with the following MANIFEST.in: recursive-include flockdoc/templates *.html In development, I'm including these (for Jinja) by doing path calculations, assuming that the "templates" directory is next to a certain file. When the package is installed with setup.py (using setuptools) the templates aren't copied into site-packages with the code. I understand that they're supposed to be somewhere like dist-packages, but none of the documentation I can find is pointing me to where the actual files are. It's also not giving me "best practice" for including these in my code. Any suggestions would be welcome there. the setup.py in question So: where are my files?

    Read the article

  • Clickone + .net client profile framework + sql compact + offline installation

    - by grimmersnee
    am having issues with prerequisites while trying to publish my windows .net 3.5 application using clickonce. I am wanting my application to work offline as well as online, so I want to include the prerequisites in the installation and not make client download them via the internet. My prerequisites are: .Net Framework Client Profile SQL Compact 3.5 I have downloaded the .Net Framework Client Profile Offline Installer. Installed it and put the DotNetFx35Client.exe in this location: C:\Program Files\Microsoft SDKs\Windows\v6.0A\Bootstrapper\Packages\DotNetFx35Client Under the Project - Publish Tab I have checked the "Download prerequisites from the following location" and entered \MachineName\Program Files\Microsoft SDKs\Windows\v6.0A\Bootstrapper\Packages\DotNetFx35Client Following http://stackoverflow.com/questions/1046370?tab=oldest#tab-top I am however still the error: The install location for prerequisites has not been set to 'component vendo's website' file Dotnetfx35client\Dotnetfx35clientSteup.exe' in item .net framework client profile can not be located on disk.

    Read the article

  • Python in AWS Elastic Beasntalk: Private package dependencies

    - by Adam Matan
    I would like to deploy a Python Flask application on beanstalk. The application depends on external packages (e.g. geopy) and internal packages (e.g. adam_geography). The manual Create a requirements.txt file and place it in the top-level directory of your source bundle. This would probably fetch geopy and its dependencies, but would not fetch adam_geography which is available from a custom repo inside my VPC. How do I specify/upload private, internal Python package dependencies in a Beanstalk application?

    Read the article

  • Why are static imports of static methods with same names legal?

    - by user1055638
    Lets say we have these packages and classes: package p1; public class A1 { public static void a() {} } package p2; public class A1 { public static void a() {} } package p3; import static p1.A1.a; import static p2.A1.a; public class A1 { public static void test() { } } I am wondering, why the static import of methods is legal (won't result in compile time error) in package p3? We won't be able to use them further in the test() method as such usage will result in the compile time error. Why it is not the same as with a normal import of classes. Lets say we would like to import classes A1 from packages p1 and p2 into p3: package p3; import p1.A1; import p2.A1; such import is illegal and will result in the compile time error.

    Read the article

  • PL/SQL exception and Java programs

    - by edwards
    Hi Business logic is coded in pl/sql packages procedures and functions. Java programs call pl/sql packages procedures and functions to do database work. pl/sql programs store exceptions into Oracle tables whenever an exception is raised. How would my java programs get the exceptions since the exception instead of being propagated from pl/sql to java is getting persisted to a oracle table and the procs/functions just return 1 or 0. Sorry folks i should have added this constraint much earlier and avoided this confusion. As with many legacy projects we don't have the freedom to modify the stored procedures.

    Read the article

  • What is __path__ useful for?

    - by Jason Baker
    I had never noticed the __path__ attribute that gets defined on some of my packages before today. According to the documentation: Packages support one more special attribute, __path__. This is initialized to be a list containing the name of the directory holding the package’s __init__.py before the code in that file is executed. This variable can be modified; doing so affects future searches for modules and subpackages contained in the package. While this feature is not often needed, it can be used to extend the set of modules found in a package. Could somebody explain to me what exactly this means and why I would ever want to use it?

    Read the article

  • How to install mod_dav_svn on Centos 5.5

    - by Min2liz
    I'm trying to install mod_dav_svn on my server (Centos 5.5, Apache Apache 2.2.11, DirectAdmin 1.35.1 ), but no luck. I'm using this tutorial: http://wiki.centos.org/HowTos/Subversion#head-ae2d6fa671ad7ebd5d7835c6edbcd15dd2d73c4d So, I'm trying in console: # yum install mod_dav_svn subversion Loaded plugins: fastestmirror Loading mirror speeds from cached hostfile * addons: centos.bst.lt * base: centos.bst.lt * extras: centos.bst.lt * updates: centos.bst.lt Excluding Packages in global exclude list Finished Setting up Install Process No package mod_dav_svn available. Package subversion-1.4.2-4.el5_3.1.i386 already installed and latest version Nothing to do Why can't it find mod_dav_svn? Also I tried: #yum search mod_dav_svn Loaded plugins: fastestmirror Loading mirror speeds from cached hostfile * addons: centos.bst.lt * base: centos.bst.lt * extras: centos.bst.lt * updates: centos.bst.lt Excluding Packages in global exclude list Finished Warning: No matches found for: mod_dav_svn No Matches found Nothing...

    Read the article

  • MVC route with id and sub-action

    - by Dan Revell
    I can't figure out what I need to do with MVC routing to make this work Here's my one route: routes.MapHttpRoute( name: "DefaultApi", routeTemplate: "{controller}/{id}/{action}", defaults: new { id = RouteParameter.Optional, action = RouteParameter.Optional } ); The request /Shipments/ works great. The request /Shipments/3/Packages works great. The request /Shipments/3 however fails with the error: Multiple actions were found that match the request: System.Linq.IQueryable`1[Api.Controllers.RequisitionsController+PackageRequisitionWithTracking] GetPackageRequisitions(Int32) on type Api.Controllers.RequisitionsController Api.Models.ShipmentRequisition GetShipmentRequisitions(Int32) on type Api.Controllers.RequisitionsController It can't seem to differentiate between: public ShipmentRequisition GetShipmentRequisitions(int id) and [ActionName("Packages")] public IQueryable<PackageRequisitionWithTracking> GetPackageRequisitions(int id) I would have thought the lack of action name on the get shipment by id would allow that route to work.

    Read the article

  • dependson option does not work in knitr?

    - by umair durrani
    Data File Name of the data file = toto.rmd Contains following: ##1st ```{r clock, cache=TRUE} x <- 600 x ``` ##2nd ```{r, cache=TRUE, cache.path="toto_cache/", dependson="clock"} x+5 ``` Problem The second chunk is not being updated. Previously x was 500, after updating it to 600, I get following after knit HTML in RSTUDIO: 1st x <- 600 x ## [1] 600 2nd x+5 ## [1] 505 What am I missing here? Session Info > sessionInfo() R version 3.0.3 (2014-03-06) Platform: x86_64-w64-mingw32/x64 (64-bit) locale: [1] LC_COLLATE=English_United States.1252 LC_CTYPE=English_United States.1252 [3] LC_MONETARY=English_United States.1252 LC_NUMERIC=C [5] LC_TIME=English_United States.1252 attached base packages: [1] stats graphics grDevices utils datasets methods base other attached packages: [1] ggplot2_0.9.3.1 loaded via a namespace (and not attached): [1] colorspace_1.2-3 dichromat_2.0-0 digest_0.6.4 evaluate_0.5.5 [5] formatR_0.10 grid_3.0.3 gtable_0.1.2 htmltools_0.2.4 [9] knitr_1.6 labeling_0.2 MASS_7.3-29 munsell_0.4.2 [13] plyr_1.8.1 proto_0.3-10 RColorBrewer_1.0-5 Rcpp_0.11.0 [17] reshape2_1.2.2 rmarkdown_0.2.64 scales_0.2.3 stringr_0.6.2 [21] tools_3.0.3 yaml_2.1.10 >

    Read the article

  • % confuses python raw sql query

    - by Jonathan
    Following this SO question, I'm trying to "truncate" all tables related to a certain django application using the following raw sql commands in python: cursor.execute("set foreign_key_checks = 0") cursor.execute("select concat('truncate table ',table_schema,'.',table_name,';') as sql_stmt from information_schema.tables where table_schema = 'my_db' and table_type = 'base table' AND table_name LIKE 'some_prefix%'") for sql in [sql[0] for sql in cursor.fetchall()]: cursor.execute(sql) cursor.execute("set foreign_key_checks = 1") Alas I receive the following error: C:\dev\my_project>my_script.py Traceback (most recent call last): File "C:\dev\my_project\my_script.py", line 295, in <module> cursor.execute(r"select concat('truncate table ',table_schema,'.',table_name,';') as sql_stmt from information_schema.tables where table_schema = 'my_db' and table_type = 'base table' AND table_name LIKE 'some_prefix%'") File "C:\Python26\lib\site-packages\django\db\backends\util.py", line 18, in execute sql = self.db.ops.last_executed_query(self.cursor, sql, params) File "C:\Python26\lib\site-packages\django\db\backends\__init__.py", line 216, in last_executed_query return smart_unicode(sql) % u_params TypeError: not enough arguments for format string Is the % in the LIKE making trouble? How can I workaround it?

    Read the article

  • Nuget and Source Control Files to Exclude?

    - by Peter Kellner
    I know I should be using nuget more but at this point I don't completely understand the nuances so I still tend to either get source and build the project, then reference the project, or I create my own "dlls" folder and hand copy the dll's in. As part of my learning process, I'm trying to understand what is critical and what is not when using nuget. For example, I've done install-package restsharp and now when I check into source control, I get files like "packages/RestSharp.103.4/lib/net4/RestSharp.xml". I'm assuming that nuget will help me with upgrading and such and it needs to have certain meta data type files. My question is: Should I be ignoring any or all files in the "packages" directory? If so, what and why. Thanks

    Read the article

  • How to build custom sqlite with provider under Android?

    - by dr4cul4
    Hi, I have a really strange problem. I need to build custom sqlite3 database engine under Android OS, but I also want to use database provider implementation. Unfortunately when examining sources of Android 1.6 I noticed that it's not so easy. Many classes including android.database.; packages use original provider, also many other parts of framework use android.database.sqlite.; packages directly, wich ofcourse make this abstraction a bit confusing and unnesesary. But going to my question. If there is any way that I could extend database interfaces to use custom implmentation of sqlite (or any other database)?

    Read the article

< Previous Page | 76 77 78 79 80 81 82 83 84 85 86 87  | Next Page >