I have gone through many documents where indices are created but the APIs used are SQL API for Azure cosmos but no example for Gremlin API. Any help would be appreciated. Thanks!!
Azure Cosmos DB is a schema-agnostic database that allows you to iterate on your application without having to deal with schema or index management. By default, Azure Cosmos DB automatically indexes every property for all items in your container without having to define any schema or configure secondary indexes.
For more details, we can see Indexing policies in Azure Cosmos DB.
Azure Cosmos DB support Gremlin API. You can use the indices for your Gremlin queries.
Reference: Querying with indexes
Hope this helps.
Related
I want to create autosuggestion within Azure Cosmos Db.
I can see that there is Azure cognitive search and Elastic stack services that can be used to achieve this. However, I have couple 1000 records only, and they are not expected to grow very large.
Now Cognitive search and Elastic stack feels like overkill.
So How can I create autosuggestion (index and queries) with Azure cosmos DB?
The Source of my NOSQL Cosmos DB data is SQL DB. I switched from SQL to NOSQL only for autosuggestion thing only because FullText search and LIKE on SQL side are not faster. dotnet core api is responsible for syncing of SQL to NOSQL Db Data.
I am new to Azure and have no prior experience or knowledge regarding working with Azure data warehouse systems (now Azure Synapse Analytics Framework)
I have access to a "read only" data warehouse (not in Azure) that looks like this:
I want to replicate this data warehouse as it is on Azure cloud. Can anyone point me to the right direction (video tutorials or documentation) and the number of steps involved in this process? There are around 40 databases in this warehouse. And what if I wanted to replicated only specific ones?
We can't do that you only have the read only permisson. No matter which data warehouse, we all need the server admin or database owner permission to do the database replicate.
You can easily get this from the all documents relate to the database backup/migrate/replicate, for example: https://learn.microsoft.com/en-us/sql/t-sql/statements/backup-transact-sql?view=sql-server-ver15#permissions,
If you have enough permission then you can to that. But for Azure SQL datawarehouse, now we called SQL pool (formerly SQL DW), we can't replicate other from on-premise datawarehouse to Azure directly.
The official document provide a way import the data into to Azure SQL pool((formerly SQL DW)):
Once your dedicated SQL pool is created, you can import big data with
simple PolyBase T-SQL queries, and then use the power of the
distributed query engine to run high-performance analytics.
You also could use other ETL tool to achieve the data migration from on-premise datawarehouse to Azure. For example using Data Factory, combine these two tutorials:
Copy data to and from SQL Server by using Azure Data Factory
Copy and transform data in Azure Synapse Analytics by using Azure
Data Factory
How to update or delete data in azure sql DB using azure stream analytics
Currently, Azure Stream Analytics (ASA) only supports inserting (appending) rows to SQL outputs (Azure SQL Databases, and Azure Synapse Analytics).
You should consider to use workarounds to enable UPDATE, UPSERT, or MERGE on SQL databases, with Azure Functions as the intermediary layer.
You can find more information about such workarounds in this MS article.
Firstly, we need to know what is Azure Stream Analytics.
An Azure Stream Analytics job consists of an input, query, and an output. Stream Analytics ingests data from Azure Event Hubs, Azure IoT Hub, or Azure Blob Storage. The query, which is based on SQL query language, can be used to easily filter, sort, aggregate, and join streaming data over a period of time. You can also extend this SQL language with JavaScript and C# user defined functions (UDFs). You can easily adjust the event ordering options and duration of time windows when preforming aggregation operations through simple language constructs and/or configurations.
Azure Stream Analytics now natively supports Azure SQL Database as a source of reference data input. Developers can author a query to extract the dataset from Azure SQL Database, and configure a refresh interval for scenarios that require slowly changing reference datasets.
That means that you can not insert or update data in azure sql DB using Azure Stream Analytics.
Azure Stream Analytics is not a database manage tool.
Hope this helps.
We are planning to use Azure Cosmos as a DB for Storing a bunch of Documents (JSON Objects), Does Cosmos have any feature to create Views like Mongodb
I searched many places but could not get details
No CosmosDB does not have, it looks like you will have to create another collection.
If you look at the commands, you can see that there is no command to create a view in CosmosDB.
Based on the REST API and Administration Commands, the feature is not supported by cosmos db mongo api yet.(Please see a similar thread:Is it possible to use MongoDB Views with Azure CosmosDB?)
Per my knowledge, CosmosDB just supports a subset of the MongoDB API. CosmosDB has some different behaviours and results. But the onus is on CosmosDB to improve their emulation of MongoDB.
In addition, you could add feedback here to submit your requirements.Or you could consider using MongoDB Atlas on Azure if you'd like full MongoDB feature support.
I've been doing some reading about sharding SQL Azure databases with Azure Elastic Scale. The article discusses two types of shard maps: List and Range. What about Hash shard maps? My application requirements are to evenly distribute data across shards which would seem to require a hashing of the shard key. Is this possible with SQL Azure and Azure Elastic Scale?
At this point in time Elastic Database tools (formerly known as Elastic Scale) doesn't support hash shard maps.
Thanks
Silvia Doomra