Background Information
TFS 2015 RC2
Release Management Server 2015
Azure VM with 2015 deployment agent
Physical local machine with 2015 deployment agent
Both machines need the drop location using the Through Release Management Server over HTTP(S) option. Currently we are using the HTTP side of things over port 1000.
Workflow
Stop App Pool (Working)
Stop Website (Working)
Copy website directory to backup location (Working)
Backup Database (Working)
Deploy Component (Not Working), using either
xcopy
msdeploy (web deploy package)
The Error (TL;DR)
The same error is received every time, it doesn't matter which machine or which deployment method. The component always fails with a JSON.NET issue.
7/22/2015 3:03:39 PM - Error - (13704, 104) - Cannot deserialize the current JSON object (e.g. {"name":"value"}) into type 'System.String[]' because the type requires a JSON array (e.g. [1,2,3]) to deserialize correctly.
To fix this error either change the JSON to a JSON array (e.g. [1,2,3]) or change the deserialized type so that it is a normal .NET type (e.g. not a primitive type like integer, not a collection type like an array or List<T>) that can be deserialized from a JSON object. JsonObjectAttribute can also be added to the type to force it to deserialize from a JSON object.
Path 'ErrorMessage', line 1, position 16.: \r\n\r\n at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.CreateObject(JsonReader reader, Type objectType, JsonContract contract, JsonProperty member, JsonContainerContract containerContract, JsonProperty containerMember, Object existingValue)
at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.CreateValueInternal(JsonReader reader, Type objectType, JsonContract contract, JsonProperty member, JsonContainerContract containerContract, JsonProperty containerMember, Object existingValue)
at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.Deserialize(JsonReader reader, Type objectType, Boolean checkAdditionalContent)
at Newtonsoft.Json.JsonSerializer.DeserializeInternal(JsonReader reader, Type objectType)
at Newtonsoft.Json.JsonConvert.DeserializeObject(String value, Type type, JsonSerializerSettings settings)
at Newtonsoft.Json.JsonConvert.DeserializeObject[T](String value, JsonSerializerSettings settings)
at Microsoft.TeamFoundation.Release.Data.Proxy.RestProxy.BaseDeploymentControllerServiceProxy.GetPackageFileInfos(String packageLocation)
at Microsoft.TeamFoundation.Release.DeploymentAgent.Services.Deployer.HttpPackageDownloader.CopyPackageAndUnpackIt(String packageSourceLocation, String filesDestinationLocation)
at Microsoft.TeamFoundation.Release.DeploymentAgent.Services.Deployer.ComponentProcessor.CopyComponentFiles()
at Microsoft.TeamFoundation.Release.DeploymentAgent.Services.Deployer.ComponentProcessor.DeployComponent()
Update (Workaround)
As a work around if I edit the build configuration to have a UNC path as the drop location, the deployment is successful. However I want to use the Copy build output to server option.
Un-installing deployer and installing RM 2015 RTM deployer should fix this issue.
There was an issue in previous RM releases where NewtonSoft.Json dll was not getting upgraded during deployer auto upgrade.
I dont think MS really tested the agent releases with Update 1.
I got the same error which is actually just a generic error message when using deployment through HTTP. When I converted it to deployment through UNC paths I found out what the problem was.
As you might know with TFS 2015 you had to name the release components exactly as the artifact names. So artifact 'WebApp X' has a release component called 'WebApp X' in RM with the subpath 'WebApp X'.
In my release configuration I have 3 different components (and artifacts).
So on the disk it was:
'\build\WebApp X'
'\build\WebApp Y'
'\build\WebApp Z'
Worked perfectly with 2015 RTM.
Now after Update 1 it looks for the following:
'\build\WebApp X\WebApp X'
'\build\WebApp X\WebApp Y\'
'\build'WebApp X\WebApp Z'
I dont know why it does this and how to solve this yet, but I manually altered the folders in the artifacts drop location and RM picked it up fine. So still looking how to fix this that it works correctly.
This json issue occurs if the user under which the RM server app pool is running doesnt have access to the drop location of the component and you have selected 'Through RM server http(s) option'
So as a fix, you can give the app pool user permissions to access the drop.
you can see the correct error in the server logs.
"Package location '\share\' does not exists or Application Pool user does not have access"
Related
I am working in a SageMaker environment that is locked down. For example, my user account is prevented from creating S3 buckets. But, I can successfully run vanilla ML training jobs by passing in role=get_execution_role to an instance of the Estimator class when using an out-of-the-box algorithm such as XGBoost.
Now, I'm trying to use an algorithm (LightBGM) that is only available via the JumpStart feature in SageMaker, but I can't get it to work. When I try to retrieve an image URI via image_uris.retrieve(), it returns the following error:
ClientError: An error occurred (AccessDenied) when calling the GetObject operation: Access Denied.
This makes some sense to me if my user permissions are being used when creating an object. But what I want to do is specify another role - like the one returned from get_execution_role - to perform these tasks.
Is that possible? Is there another work-around available? How can I see which role is being used?
Thanks,
When I encountered this issue, it was a permissions issue with a bucket that had changed.
In the SageMaker Python SDK source code , there is a cache that is located at in an AWS-owned bucket: jumpstart-cache-prod-{region}. and a manifest.json that translates the ECR path for the image for you.
If you look at the stack trace, it could be erroring out at the code that is looking for the manifest.
One place to look is if there are new restrictions placed in IAM, Included here is the minimum policy you need to access JumpStart (pretrained) models
I have created an ISE Logic Apps environment, and am trying to add the FileSystem ISE connector to the Managed connectors list, as it does not appear in my default list. When I click on '+Add', I can see the File System connector in the drop down that comes up, but when I select it & click on 'Create', I get the following error:
Failed to create connector 'isefilesystem'
Operation name
Set Integration Service Environment managed API
Time stamp
Mon Jan 13 2020 16:53:24 GMT+0000 (GMT Standard Time)
Event initiated by
xxxxxxxxx#xxxx.xxx
Error code
IntergrationServiceEnvironmentManagedApiDefinitionTagsNotSupported
Message
The tags are not supported in the managed API 'isefilesystem'.
The ISE File System connector is available. The ISE File System connector will not be automatically added to the ISE, you will need to manually add it yourself. Our engineers are working on getting that automatically added with new ISE deployments.
Here are steps from the document to add it manually to the ISE :
On your ISE menu, under Settings, select Managed connectors. On the toolbar, select Add.
On the Add a new managed connector pane, open the Find connector list. Select the ISE connector that you want to use but isn't yet deployed in your ISE. Select Create.
Only ISE connectors that are eligible but not yet deployed to your ISE appear available for you to select. Connectors that are already deployed in your ISE appear unavailable for selection.
Checkout this GitHub issue as well for details.
File system connectors is not yet available in ISE. You can use the shared connector (With gateway) in a Logic app in ISE till it is available.
It is on-premises data gateway. Yes there is already a work items for this and it is in progress and will be there in near future but this timeline is subject to change. As far as new features go, we are not able to disclose much at this time.
Also you could vote up this feedback to promote this feature to be achieved quickly.
I am having problems with the ISE Create File (Preview).
I have an API Connection defined with the Root Folder setting using dot notation, e.g. 192.168.1.23, because there is a DNS issue with hybrid cloud-on-prem lookup, or so I am told.
The Logic App portal editor in Designer mode behaves strangely when configuring the folder path in the Create File action. When using the pop-out folder picker I see "The use name or password is incorrect".
I have made sure that the credentials are correct and have tested successfully via other means.
Is there a work around?
Is this a known issue?
I would like to host a tensorflow model (trained on my local PC) on Sagemaker and I followed this article: https://aws.amazon.com/fr/blogs/machine-learning/bring-your-own-pre-trained-mxnet-or-tensorflow-models-into-amazon-sagemaker/
However I'm not able to perform the inference. My model is an object detection model, locally it has generated frozen_inference_graph.pb, model.ckpt.data-00000-of-00001, model.ckpt.index and model.ckpt.meta, all this files have been converted into save_model.pb and variables.data-00000-of-00001, variable.index, then into .tar.gz by respecting the folder structure export/Servo/version/...
.tar.gz has been manually uploaded to s3 and I have successfully created an endpoint.
But when I try to perform inference, I have an error:
ModelError: An error occurred (ModelError) when calling the InvokeEndpoint operation: Received client error (415) from model with message "{"error": "Unsupported Media Type: Unknown"}"
My input datas are images. I need help.
My new client has an unmanaged package in their production org. I have access to the package / source code, but I can't modify this instance of the code. I need to move a copy of the package into my development org. I'm attempting to use the Force.com IDE to deploy the package, but I'm getting a series of errors such as the ones below.
I'm new to Salesforce development. Any help would be appreciated
* Deployment Log *
Result: FAILED
Date: August 17, 2014 9:52:04 AM EDT
Deployed From:
Project name: xxx
Username: xxx
Endpoint: www.salesforce.com
Deployed To:
Username: xxx
Endpoint: www.salesforce.com
Deploy Results:
File Name: classes/CreateNewVoucherController.cls
Full Name: CreateNewVoucherController
Action: NO ACTION
Result: FAILED
Problem: sObject type 'Kumo_Vouchers_Group__c' is not supported. If you are attempting to use a custom object, be sure to append the '__c' after the entity name. Please reference your WSDL or the describe call for the appropriate names.
File Name: classes/GatewayAuthorizeNet.cls
Full Name: GatewayAuthorizeNet
Action: NO ACTION
Result: FAILED
Problem: Invalid type: GatewayBase
File Name: classes/GatewayBase.cls
Full Name: GatewayBase
Action: NO ACTION
Result: FAILED
Problem: sObject type 'Order__c' is not supported. If you are attempting to use a custom object, be sure to append the '__c' after the entity name. Please reference your WSDL or the describe call for the appropriate names.
based on provided error description I can assume that you tried to deploy a code, but missed to deploy a sObjects from package.
Try to deploy in a few steps:
deploy all sObjects (each object might has an dependency to another object, thus you have to deploy objects without dependencies). Also objects might have dependencies on VF pages which can't be deployed without apex, which in its turn can't be deployed without object. Thus you can comment out a part of sObjects with relates to VF page and uncomment after VF-page deployment.
deploy all layouts, object translation, custom labels, applications, static resources, workflow rules
deploy apex code
deploy apex triggers & VF pages
deploy all other components.
In our team, we use a common database server. When running the Orchard Setup, I point the connection string to an existing Orchard Database and got this error message:
Setup failed: The requested service 'Orchard.IWorkContextAccessor' has not been registered. To avoid this exception, either register a component to provide the service, check for service registration using IsRegistered(), or use the ResolveOptional() method to resolve an optional dependency.
The only way that can make setup successful for me is to point to an empty database. I attempted a workaround by doing so first, and then edit connection string in App_Data\Sites\Default\Settings.txt, but then I got the same error.
I had this problem as well, also tried copying the settings.txt file which did not work. This time I copied all the contents of the App_data directory over and it worked for me. Also, this latest time I was using the latest commit from codeplex.
I deleted cache.dat and restarted the website in IIS, this fixed it for me