I'm working on a connector. I have the following error when I'm trying to deploy my script:
There was an error retrieving this Add-on's configuration.
What does it mean?
It's hard to say without seeing the code / knowing what you are trying to do.
Several possibilities:
getConfig is not defined or returns wrong values.
getConfig has an error.
Related
data training code error
I wanted to solve this for data training. But its giving this error and I don't know how to solve. Tried to solve it via many methods but still stuck on this.
I tried to rename it but couldn't get it solved.
I have created an Ontology and have run the reasoner successfully. However upon adding a data property called isLocatedExternally and associating it with individual x
and running the reasoner I get the following error
MalformedLiteralException: Literal"C:\Program Files"^^<http://www.w3.org/2001/XMLSchema#anyURI>is malformed.
I am trying to assign the individual to the file being stored locally on my PC.
Is there any fundamental step I am missing?
I had initially had my file on the cloud and thought storing the file on my PC would help but that did not change anything.
Below is the screenshot of the error and the data property created.
What I am trying to do is to use the technique below for assigning an external source to an individual.
So far, based on the comments, I have tried
1- file:///C:\Program%20Files
2- file://C:\Program%20Files
3- File://C:\Program%20Files
4- C:\Program%20Files
Here is the reasoning error that arises when I run the reasoner when assigning the individual to an external file using anyURI
file:///C/Program%20Files resolved the issue. However this raises another issue which is when I change to this address, and lookup for this address in the file explorer in the windows OS, it cannot find the path and gives the following error:
Windows cannot find "file:///C/Program%20Files".Check the spelling
and try again
Now the IRI does not direct the individual to an existing folder
Based on my personal experience you need to retry with other reasoners and see if that helps
The problem I am trying to export my redcap data to a CSV form, and unable to do so. I am receiving an error notification that says there is to much data, although it is a tiny project. help will be much appreciated.
The full error text: " We are sorry, but apparently the data export is not able to complete successfully. It may simply be that there is too much data trying to be exported at once, in which it is causing REDCap to crash. If this error occurs again, it is recommended that you attempt to export a smaller data set (fewer fields and/or perhaps fewer records) so that this error does not occur. Our apologies for this inconvenience."
what I have tried
I have made sure I have the necessary user rights.
I have tried through a colleges redcap user (who has thenecessary user rights) .
have tried exporting only one instrument (no success)
have created a test project with only 2 questions. also in the new test project I receive the same notification.
could not export data both in development mode and in production.
any ideas?
Many Thanks!
The institution had blocked the option for file upload due to security reasons. Apparently REDCap's exporting system uses the uploading mechanism, and therefore landed up being disabled as well.
The local storage folder location was pointing to a folder that was missing on the server. Just created the folder and upload and subsequent problem was fixed
I'm using evolution scripts, via Slick, using the Play Framework to update the schema for a Microsoft SQL Server database.
This largely works great, except that when something goes wrong, I just get a terse error message, with no indication which line of the script caused the error, and—for large scripts—that makes identifying the error time-consuming and challenging.
For example, consider this error:
An evolution has not been applied properly. Please check the problem and resolve it manually before marking it as resolved. -
We got the following error: , while trying to run this SQL script:
That's it.
Is there a way to get the full error message (through a log file, a configuration setting, etc.) that includes the line number and context, etc. with the evolution failure?
It took me a while, but I finally figured this out.
I edited the conf/logback.xml file to include the following:
<logger name="play.api.db.evolutions" level="DEBUG" />
Each statement in the Ups script is then written to logs/application.log as it is executed, together with its result, including detailed error messages.
I am trying to import a CSV file into my BigQuery Table. This import has worked in the past, but now I am getting the following error message:
{"message":"Too many errors encountered. Limit is: 0.","reason":"invalid"}
All other fields are empty when I run the debugger.
This is... not helpful. I am unaware of any issues with the data itself, as the export/import data has not changed. Curiously, when trying to use a previous Job Template and run through the web console, the web console itself hangs and the dialog never goes away once I hit the blue "Submit" button.
Job Id: job_e0faf560d3df424ea74519e1b24a23f7
I am generating a CSV and exporting it to Google Cloud Storage. I am using AppEngine and have switched to the new Google Cloud Storage Client Library. I had uploaded the file using the GcsFileOptions.getDefaultInstance() as well as constructing my own GSFileOptions setting the content type to CSV.
After failure, I downloaded the file from Google Cloud Storage, change the encoding (tried ASCII and UTF8) and still have gotten the same result.
I am using AppEngine 1.8.1.1 and the BigQuery Library (google-api-services-bigquery-v2-rev89-1.15.0-rc). This was working as expected previously, so I'm not sure what has happened. Any suggestions are welcome. Thank you!
There are two error fields on the bigquery job. The first is the error result, which tells you whether (and why) the job failed. The error result in your case that the job failed due to encountering too many input errors during the import.
The second field is the error stream, which tells you about errors encountered during the job. If you had set the maxBadRecords field, for example, you could have errors in the error stream, but the actual job might succeed.
I looked up your job in the BigQuery logs, and was able to find that the error stream indicates an error on line 6253: "Too few columns: expected 80 column(s) but got 1 column(s). For additional help: http://goo.gl/RWuPQ"
Can you verify that line 6253 is correct?
-- Jordan Tigani / BigQuery Engineer
Today there is some general problem with app engine:
"We are still investigating the issue with Google App Engine, primarily (but not restricted to) Datastore latency.
We will provide another status update in the next two hours."
https://groups.google.com/forum/#!topic/google-appengine-downtime-notify/1pJZnl4EMKk