I can't clone git or upload my web to gcloud and the only one notice I get is
"error: unable to create temporary file: No space left on device"
What I've tried to do:
1. Restart VM
2. Expanse storage
And its not solved yet. Pls help
Related
I am using VS code to open my react projects. Neither the folder from the git repository nor the locally stowed projects are accessible.
[13512:1219/213241.168:ERROR:cache_util_win.cc(20)] Unable to move the cache: Zugriff verweigert (0x5)
[13512:1219/213241.169:ERROR:cache_util.cc(145)] Unable to move cache folder C:\Users\magsu\AppData\Roaming\Code\Cache\Cache_Data to C:\Users\magsu\AppData\Roaming\Code\Cache\old_Cache_Data_000
[13512:1219/213241.169:ERROR:disk_cache.cc(196)] Unable to create cache
I tried to reboot, cleaned and emptied the cache, deinstalled and reinstalled VS code.
I have experienced the same error message with node and python applications. In my case the issue was that I have executed the script file without the language prefix. Instead of:
node myscript.js
python myscript.py
I have executed:
myscript.js
myscript.py
I hope that this answer helps anyone having the same issue.
Either the permissions are bad or some process is holding a lock on those files (more likely).
I'm guessing in the background you have some process open that uses these files. For example, a dev server. Try killing all node processes via Windows Task Manager if you can't find the window and then try again.
i'm new to Openshift.
Here my problem:
i've deployed a war, using git and it works fine. I have a servlet that upload images, and everything is okay.
I don't understand, where should I go to see the structure file system.
For example, in Tomcat in eclipse I can see the uploaded file in the file system.
Is there the chance in openshift to see my file system and so my uploaded file?
And more, if I deploy a war, may I have the chance to modify it with some kind of console in openshift?
Thanks for answering.
The right way to do this is using rhc client.
So these are the steps to follow:
1) install ruby-installer
2) install git
3) install rhc
4) type rhc setup.
Then you will be asked to input your credentials. Once logged in, you will obtain an OAuth token, and a new public key is uploaded to openshift.
Then type:
rhc -a app_nome and you will have the app view.
Then, typing ls, you will see the entire file system.
I'm running the following command to deploy my Managed VMs app (on Windows 10):
gcloud preview app deploy app.yaml --project=<PROJECT> --promote
The deployment starts bug hangs on the following line:
Copying certificates for secure access. You may be prompted to create an SSH keypair.
And after some time I get the error:
ERROR: (gcloud.preview.app.deploy) Unable to copy certificates.
I've already:
Made sure that there are SSH keys in ~\.ssh\google_compute_engine
Tried to run with --quiet - same results
Renamed ssh-term.exe to ssh.exe - same results
Run the command as an administrator.
Run the command with --verbosity debug, which prints the following line multiple times: DEBUG: File [f] does not exist locally.
Any help will be much appreciated!
Found the cause! It was the project's firewall that blocked SSH by default. Fixed that and it worked.
Glad you fixed it, I had the same problem and will use your fix. I did happen accros a work around. By using the Container Build API to perform the build.
enter the command
gcloud config set app/use_cloud_build true
Before you
gcloud preview app deploy
Cite: https://github.com/isusanin/google-cloud-sdk/issues/533
There is more than one example where I have seen this status from a pod running in openshift origin. In this case it was the quickstart for the cdi camel example. I was able to successfully build and run it locally (non - openshift) but when I try to deploy on my local openshift (using mvn -Pf8-local-deploy), I get this output for that particular example (snipped for relevance) :-
[vagrant#vagrant camel]$ oc get pods
NAME READY STATUS RESTARTS AGE
cdi-camel-z4czs 0/1 CrashLoopBackOff 4 2m
Tail of the logs are as under:-
Error occurred during initialization of VM
Error opening zip file or JAR manifest missing : agents/jolokia.jar
agent library failed to init: instrument
Can someone help me solve this ?
If the state of the pod goes in to CrashLoopBackOff it usually indicates that the application within the container is failing to start up properly and the container is exiting straight away as a result.
If you use oc logs on the pod name, you may not see anything useful though as it would capture what the latest attempt to start it up is doing and may miss messages.
What you should do is instead provide the --previous or -p option to oc logs along with the pod name. That will show you the complete logs from the previous attempt to start up the container.
If this is an arbitrary Docker image you are using, a common problem that can occur and which would cause the container not to start, is an application image which requires to be run as the root user. Because running an application inside of a container as root still has risks, OpenShift doesn't allow you to do that by default and will instead run as an arbitrary assigned user ID. The application image may not be designed with this possibility in mind and so is failing.
So try and get those logs messages and see what the problem is.
Temporary workaround -> https://github.com/fabric8io/ipaas-quickstarts/issues/1157
Basically, the src/main/hawt-app directory needs to be deleted.
I needed to test some changes on my local dev server before pushing to production. Doing so required having the full dataset on my local machine.
A colleague directed me to:
https://developers.google.com/appengine/docs/python/tools/uploadingdata?csw=1
I downloaded the data using an administrator's username and password, but unfortunately, I was unable to upload the data to my localhost "dev" app engine server.
Ran this command from the commandline:
appcfg.py upload_data --filename=../data/data1.dat --url=http://localhost:9080/_ah/remote_api ./
Where:
9080 was my app port on my localhost copy of the app
I was running this command from my app directory
Had the downloaded data stored in relative directory
../data/data1.dat
Received this error:
raise _ToDatastoreError(err)
google.appengine.api.datastore_errors.BadRequestError: app "dev~appname" cannot access app "appname"'s data
UPDATE: It seems that the answer was as simple as adding the following to my upload_data call:
--application="dev~appname"
Thanks #DavidBennett.
ORIGINAL ANSWER: (which also works)
After a ton of searching on SO and code.google.com, the solution I found that worked was a comment on this question:
devappserver2, remote_api, and --default_partition
I used my original command as described in the question:
appcfg.py upload_data --filename=../data/data1.dat --url=http://localhost:9080/_ah/remote_api ./
The username and password I entered when prompted were my apps username (in my case, my email) and the corresponding password. (If that doesn't work you might want to try blank or test#example.com based on other comments I've read, but have not tested that theory.)
I also restarted my app engine with the following flag: (don’t forget to remove the flag the next time you restart the server) (You might want to try without using this flag, since I can’t confirm that it affects anything - I’m including it here, since it was a setting that I used.)
--clear_datastore=yes
The commenter recommends to delete “dev~” in your local server code on line 84 in this file:
google/appengine/tools/devappserver2/application_configuration.py, line 84
Where:
that base directory 'google' is located inside of:
/Applications/GoogleAppEngineLauncher.app/Contents/Resources/GoogleAppEngine-default.bundle/Contents/Resources/google_appengine/
assuming your GoogleAppEngineLauncher.app directory is in your Applications directory on your Mac
IMPORTANT: Restart your local app engine server for the changes to take effect.