Deployment failed case on Devtestlab - azure-devtest-labs

I am trying to add a claimable VM for my lab in Azure DevTest following this document Create and manage claimable VMs - Azure DevTest Labs | Microsoft Learn but whenever I try to add
Operating system = windows
base vm = windows server 2022 datacenter azure editon core
size = standard_B1s
I am getting failed yet each time I attempt to add
And the disk showing as No result
Any solution while using this particular data center it fails or am i doing something wrong and I want this particular data center to have a fix for the issue?

I tried to reproduce the same in my environment and got the same error like below:
I created Devtest lab like below:
To resolve this issue try to add Artifact to configure and add to the VM like below:
Select Artifact:
In Additionally try to set an expiration date that you might want to delete VMs automatically after a certain date so they don't keep incurring costs. To create more than one VM with this configuration select an instancelike below:
Now, Deployment is created with disk successfully like below:


KingswaySoft SSIS error when Published - A password is required

I've created a package that gets some information from a sql database and inserts them into Dynamics CRM.
When testing the package from Visual Studio everything goes as expected and the task finishes without any errors and the rows get inserte . However , when I publish the package to SSISDB on Sql Server the package fails with this error :
KingswaySoft.IntegrationToolkit.DynamicsCrm.CrmServiceException : CRM service call returned an error : A password is required in order to establish the connection ...
I tried changing the package protection level to EncryptSensitiveWithUserKey but it still gives the same message as above , created the package again from scratch still doesn't work . This package was working before , maybe there's something I did the last time in configuration which made it work but I cannot replicate it anymore .
Also I tried the Integrated Authentication it says this :
KingswaySoft.IntegrationToolkit.DynamicsCrm.CrmServiceException : CRM service call returned an error : The caller was not authenticated by the service .
#Drinv, this is a typical SSIS runtime deployment issue. You need to make sure that you have provided a password for your job configuration for the connection manager. What you provided to the package doesn't count as far as sensitive fields are concerned (password being one) when you are using the EncryptSensitiveWithUserKey option since user key is not transferrable between different systems or different users. An easy workaround is to change your SSIS package/project's ProtectionLevel setting to encrypt using a password instead, although it may not be the best practice. If you still have trouble getting this going, please reach out to us directly, our team can walk you through the issue.
I found out what I was doing wrong .
My SSIS project was on Project Deployment Model and I was trying to deploy only the package. After making my connections available on project level and deploying the whole project everything worked as expected .

SQL Server Reporting Services HTML Error 500.24

I have been using SQL Server Reporting Services on many servers but on this particular one when I try to open the url to create subscriptions I get the following error:
HTTP Error 500.24 - Internal Server Error An ASP.NET setting has been detected that does not apply in Integrated managed pipeline mode
Any advise would be greatly appreciated, please advise if more information is required
One option would be to simply switch your Managed Pipeline mode from Integrated to Classic in IIS.
The steps would be as follows.
Start the Internet Information Services (IIS) Manager from the Start menu on the Windows machine.
Click on Application Pools from the left menu and select the application you are working on from the middle table.
Right-cick to bring up the "Set Application Defaults" menu.
Change Managed Pipeline Mode from "Integrated" to "Classic" and press OK.
You will then see the Managed Pipeline column for that item change from Integrated to Classic.
This should at least get rid of the error message, without diagnosing the cause (which would require a lot more information).
Issue resolved, I had to create the virtual directory for Report manager URL.Once I added the virtual directory name and clicked apply, the report manager opened in IE.

Azure Search not recognizing Integrated Change Tracking on SQL Server Database

I am currently setting up our second Azure Search service. I am making it identical to our existing one, just in a different region.
I'm using the portal Import Data function to set up my index. For the Data Source, I have configured it to connect to my Azure SQL Database and table, which definitely has Integrated Change Tracking turned on. Further, it's the exact same database and table that I'm connected to and indexing from in my existing Azure Search service.
The issue is that when I get to the "Create an Indexer" step, I get the message that says "Consider enabling integrated change tracking on your database..." In other words, it doesn't think I have change tracking on this database. I definitely do, and our other Azure Search Service recognizes this just fine on the exact same database.
Any idea what's going on here? How can I get this Data Source to be recognized as having Change Tracking turned on, and why isn't it doing so when all is working as expected in our existing Search service with identical set up?
We will investigate. In the meantime, please try creating your datasource and indexer programmatically using the REST API or .NET SDK.
When I was experiencing this problem, I tried creating the search service via "Add Azure Search" in Azure portal > SQL database.
Using that wizard I was able to create the search data source, index & indexer.
Update: I opened a ticket with Azure support, and when trying to get more information to provide to them, I tried to reproduce the problem (create a data source via REST API), but the expected failure did not happen ("Change tracking not enabled for table..." despite it being enabled). This makes me think there was something wrong with internal Azure code that was fixed in the meantime. - MSaccess connection fail - already installed Access Database Engine 2016

I need to make an application for my school project. The purpose of the app is getting information from a specific database and filter the desired data.
I made a test-database so I can already test it on During the connection I keep getting the error 'The Microsoft.ACE.OLEDB.12.0 provider is not registered on the local machine'.
So after getting the error , I made some research and found that I need to install Access Database Engine. However after I installed it , I keep getting the same error.
I have already changed the provider setting in to Microsoft.ACE.OLEDB.12.0.
So I would like to ask you how I can solve this problem.
Thanks in advance

How can I resolve error when trying to launch instance on Amazon RDS?

I'm using the AWS Toolkit in Visual Studio 2013 to attempt to launch a new instance on Amazon RDS. I get through the wizard for creating the new instance and after clicking finish, there is a delay, and then a message appears saying:
Error launching DB instance: DB Security Groups can only be associated with VPC DB Instances using API version 2012-01-15 through 2012-09-17.
Launching different types of instances (SQL Server SE vs MySQL) doesn't seem to help, nor does selecting different versions of the platforms (SQL Server 2008 vs 2012). The only thing that gets it to go through is unchecking the box for "default" in the DB Security Groups area. However, I feel like something is going on here that shouldn't be happening.
Can anyone explain why this is happening and how I can resolve it other than by not setting a default security group? Thank you.
If you created your AWS account recently, you will be using a VPC by default.
It sounds like the API the plugin is trying to use hasn't been updated. The latest version is 1.5.6, and looking at the history it seems like some of these features were added in 1.5.0.
I finally solved it! Since I couldn't use the API that the VS 2013 plugin uses, I had to manually add my IP to the Security Group created for my Elastic Beanstalk.
Go to the console, ec2's security groups configuration
Find the one which description matches your Beanstalk (e.g.: Security Group created for Beanstalk Environment to give access to RDS instances)
Hit Inbound, Edit and add a new rule for All Traffic (I guess HTTP should be enough, but just in case).
In Source, select My IP and Save.