WPF and Squirrel deployment on Azure - wpf

I have an application written in WPF (C#) and I deploy this application using Squirrel:
https://intellitect.com/deploying-app-squirrel/
Now I build application using these commands:
.\nuget pack nuget\HelloWorld.nuspec
Squirrel --releasify HelloWorld.1.0.0.nupkg --releaseDir "C:\SquirrelReleases"
on my local machine. But my application is stored in Azure.
How to make deployment using Squirrel on Azure?

How to make deployment using Squirrel on Azure?
Reading the documentation here, it seems there's no direct way to deploy your files in Azure Storage at least as of answering this question. Even for Amazon S3, they are mentioning that you upload the files manually in a s3 bucket:
5. upload the files from the Squirrel Releases directory into the S3 bucket.
I guess you can do something similar for Azure Storage as well. I have not tried it but I believe this is what you would need to do (based on their documentation for Amazon S3):
Create a blob container in your Azure Storage account and set its access level to either Blob (recommended) or Public.
Update the package location on the UpdateManager in your application to use the blob container URL (https://account.blob.core.windows.net/blob-container-name).
Upload the files in blob container. There are many options for you to do that from using available storage explorers, or using AzCopy, or using Azure PowerShell/CLI tools or writing code yourself by using any of the available SDKs.

Related

How to get database off of localhost and running permanently?

So not sure it this is stupid to ask, but I'm running a neo4j database server (using Apollo server) from my React Application. Currently, I run it using node in a separate terminal (and I can navigate to it on localhost), then run npm start in a different terminal to get my application going. How can I get the database just up and running always, so if customers use the product they can always access the database? Or, if this isn't good practice, how can I establish the database connection while I run my client code?
Technologies being used: ReactJS, Neo4j Database, GraphQL + urql
I tried moving the Apollo server code into the App.tsx file of my application to run it from there directly when my app is launched, but this was giving me errors. I'm not sure if this is the proper way to do it, as I think it should be abstracted out of the client code?
If you want to run your server in the cloud so that customers can access your React application you need two things:
one server/service to run your database, e.g. Neo4j AuraDB (Free/Pro) or other Cloud Marketplaces https://neo4j.com/docs/operations-manual/current/cloud-deployments/
A service to run your react application, e.g. netlify, vercel or one of the cloud providers (GCP, AWS, Azure) that you then have to configure with the server URL + credentials of your Neo4j server
You can run neo4j-admin dump --to database.dump on your local instance to create a copy of your database content and upload it to the cloud service. For 5.x the syntax is different, I think neo4j-admin database dump --path folder.

Automate the Upload of a .bak file to Azure

I'm very new to Azure and have been tasked with automating the process of taking an existing version of our database, converting it to the newer version and then uploading that to Azure.
The conversion is done, that parts easy, what I'm struggling with is getting a .bacpac file from SSMS using PowerShell. I know I can use the Export Data Tier Application function in SSMS to do this but I need it to be automated. From there I can use something like the following to actually upload the database:
https://blogs.msdn.microsoft.com/brunoterkaly/2013/09/26/how-to-export-an-on-premises-sql-server-database-to-windows-azure-storage/
I have looked around and cannot find a solution to this, or even know where to start.
You can create bacpac of your on-premises databases and locate them on a local folder (c:\MyBacpacs) using SQLPackage.
sqlpackage.exe /Action:Export /SourceServerName:. /sdn:"DB_Foo" /tf:"c:\MyBacpacs\DB_Foo.bacpac"
You can then use AzCopy to upload bacpacs to Azure BLOB storage
AzCopy /Source:"c:\MyBacpacs" /Dest:"https://exampleaccount.blob.core.windows.net/bacpacs" /DestKey:storageaccountkey /Pattern:*.bacpac

Upload Images to Azure Storage Using Portal (not programmatically)

I need a SQL Server database that stores images, and their name, category, etc, so the SQL table will have 5 or so columns. I'm using Azure as my SQL Server host. It appears I cannot seem to insert image data into my VARBINARY(MAX) column from SQL Server Management Studio which was my first plan. I cannot do this because I cannot seem to give my user permissions to use BULK LOAD. Azure SQL seems to make this impossible. I think I need to use Azure Storage, and then in the SQL Server database, just store a link to the image.
To be clear, I want the images in the database already, I do not want to add them from within the application I am developing. The application I'm developing will only download the images to the device, not upload them.
So How do I upload the images to Azure Storage using the portal, not using code?
So how do I upload the images to Azure Storage using the portal, not using code?
Short Answer
You cannot. The portal does not have a way to upload an image to a storage container from either the old or the new portal.
Alternative
Use the AzCopy Command-Line Utility by Microsoft. It allows you to do what you want with just two command lines. There is terrific tutorial here.
First, download and install the utility. Second, open a command prompt and navigate to the installation AzCopy install directory. Third, upload a file to your storage account. Here are the second and third steps.
> cd C:\Program Files (x86)\Microsoft SDKs\Azure\AzCopy
> AzCopy /Source:folder /Dest:account /DestKey:key /Pattern:file
And here are what the parameters mean.
Source The folder on your computer that contains the images to upload.
Dest The address of the storage container at which to store the images.
DestKey The primary access key for your storage account.
Pattern The name of the file to upload (or a pattern).
Example
This uploads an image named my-cat.png from the C:\temp folder on my computer to a storage contained called mvp1. If you wanted to upload all the png images in that folder, you could replace my-cat.png with *.png and it work upload them all.
AzCopy /Source:C:\temp /Dest:https://my.blob.core.windows.net/mvp1 /DestKey:tLlbC59ggDdJ+Dg== /Pattern:my-cat.png
You might also what to take a look at the answers to this question: How do I upload some file into Azure blob storage without writing my own program?

Deploying angular-fullstack to Azure

Hi I was wondering how I can deploy angular-fullstack to Azure, I have looked around but can't find any tutorials etc. Can anybody provide any instructions or examples?
https://github.com/DaftMonk/generator-angular-fullstack
My understanding is you have two problems to solve to get this into azure.
How do I best host my mongo db?
How do I publish my node.js application as an azure app?
Answers.
Deploy your mongo db either using a VM or a managed mongodb service from the marketplace.
a. Worker role option: https://github.com/mongodb/mongo-azure (worker role mongo db)
b. Managed Mongo Hosted Azure Service - You can find this from the marketplace from your azure console/portal. (Note that not all regions apply and its not cheap)
To deploy your node.js app you have two options.
a.
Create an azure app and hook up the source control so that it will deploy on every commit.
This is the step by step guide from MS https://azure.microsoft.com/en-gb/documentation/articles/cloud-services-nodejs-develop-deploy-app/
b.
To publish, run the Publish-AzureServiceProject cmdlet as follows:
Publish-AzureServiceProject -ServiceName NodeHelloWorld -Location "East US" -Launch
-ServiceName specifies the name for the deployment. This must be a unique name, otherwise the publish process will fail.
-Location specifies the datacenter that the application will be hosted in. To see a list of available datacenters, use the Get-AzureLocation
cmdlet.
-Launch opens a browser window and navigates to the hosted service after deployment has completed.
To get the full info on deploy node apps to azure check out the microsoft docs, its pretty comprehenssive with screenshots too.
https://azure.microsoft.com/en-gb/documentation/articles/cloud-services-nodejs-develop-deploy-app/

Key steps to uploading a Drupal website from Local to live using a hosting firm

I'm a newbie to pushing Drupal websites from local to live via a CP panel with a hosting company and wondered if there are any key steps I need to follow? I usually end up with Internal Server 500 errors or no themes showing so not a good start!
The steps I follow are:
Export the database from my local PHPMyAdmin
Log into my hosting CP Panel and create the database on there
Create a user for the database (with password)
Change the settings.php to match the database settings
Load all Drupal files via FTP
Create a 'tmp' folder in the 'sites > default> files' directory
What am I doing wrong?! Is it something to do with the .htaccess file as to why I either get the error or my theme never shows?
Any help would be much appreciated! So stressful and frsutrating as a newbie! Once I've done 1 I'm hoping it'll be plain sailing!!
Thanks!
C
You have the basic steps right. Check the php error logs on the server (probably accessible via the control panel if you dont have ssh access), they should give you more information as to what actually caused the 500 errors.
Doubt it is an htaccess issue unless you are doing something crazy in there.
Can you see he drupal admin at all? If so, clear cache, check watchdog for clues also.
It's easier to download and install Drupal again on the live server rather than to copy everything via FTP. The settings.php file is where your MySQL information is stored so this file should not be copied. Follow Drupal's documentation on how to install Drupal at https://drupal.org/documentation/install/download
To transfer your database, install and enable the Backup and Migrate module on your local server from https://drupal.org/project/backup_migrate and back up your database locally.
After Drupal is installed on the live server, go ahead and copy your modules, themes, and files from /sites/all and /sites/default/files (or any non-Drupal core files that you may have created). Enable and use the Backup and Migrate module to restore your database to your live server. You may need to configure the php.ini file if the database is over 8MB.

Resources