I have a situation where I am uploading an image in sharepoint and it is being saved using blob. I need to create an XML file with the data of the blob and other data that helps users to identify it. The following is a hint of what I want
<image>
<name>mydog</name>
<extension>.jpg</extension>
<blobid>0234234</blobid>
<blobpath>435343445</blobpath> </image>
I was looking at the tables in wss_content and came up to alldocumentstreams where there is a column called rbsid. unfortunately I cannot link this id to non of my documents. My question is this is there a way how i can get all the blob information from the DB so i can link it to other details?
Directly accessing the SharePoint database isn't supported by Microsoft.
If a server component requires information from the database, it must
get that data by using the appropriate items in the SharePoint object
model, and not by trying to get the items from the data structures in
the database through some query mechanism.
You might be better using the SharePoint object model to read these files.
Some links that should help
http://www.codeproject.com/KB/sharepoint/File_Shunter.aspx
http://www.learningsharepoint.com/2011/04/01/read-a-file-in-sharepoint-document-library/
Related
I am new to azure logic app and am trying to develop our first logic app in the designer. The goal is simple - connect to our salesforce production instance and extract data from an object. Write that data to a delimited file in our storage container - from where the data warehouse can ingest the file.
The first steps are pretty straight forward - i added a recurrence trigger and connected a salesforce Get Records action to it. Chose the User object in salesforce for now because its got minimal data.
Executed the logic app and it runs fine. I can see the data extracted is correct.
However now I am lost as to how to create a delimited file in the container. I obviously could not connect the Create Blob action directly to the Get Records action. Based on some documentation i tried creating a CSV table first but got an error saying that an Array is expected but the Get Records action returns an object.
If anyone has faced something similar and has any pointers, documentation would appreciate it. Thanks in advance for any help
Regards
Sid
The Get records object will have 2 objects '#odata.context' and 'value'. The 'value' field will have actual array of records, so you need to use the value.
Please refer below screenshot of the logic app that works, fetched record from user object and stored as delimited csv file.
Here is output screenshot for create csv table:
what actions/connectors one could use to parse a excel file and map columuns from a excel table to and external store, in this case i wish to make a record in Dynamics 365 operations using OData entity.
Thanks
Why exactly are the obvious RapidStart Services no option for you?
If you definitely need to use OData, I suggest building a publishable OData page via WebServices. You then can implement a parser in any environment/language you prefer and submit the new record via a RESTful Webservice-Call (which nearly every framework should allow) to this very same page, in order to submit the record to your productive environment.
I have multiple Azure tables across multiple Azure storage that have the exact same format. Is it possible to configure several data sources in Azure-search to use a unique Index so that a search on this Index would return the results aggregated from all data sources (Azure tables)?
So far, each time I configure a new 'Data Sources' and the corresponding index, I must create a new index (with a new index name). Attempting to reuse an existing index name results in an error stating "Another index with this name already exists"
Thank you for any help or pointer you might provide.
Yes, it's possible, but we don't currently support it in the Azure Portal.
When you go through the "import data" flow in the portal, it'll create a data source, indexer and index for you.
If you want more sources for that index, you need to create new data sources and indexers, with the new indexers pointing at the existing index. Unfortunately this is not currently supported from the portal. You can do it using the .NET SDK (if you're using .NET), directly using the REST API from your app, or using any tool that can make HTTP requests such as PowerShell, curl or Fiddler.
The documentation that describes the indexer-related REST APIs is here:
https://msdn.microsoft.com/en-us/library/azure/dn946891.aspx
I have a specific field that I am trying to find. The salesforce instance I am in has hundreds of tables/objects so I can't look through them manually.
I also only have read only access, so I can't run an APEX script or create objects. I am using an API to access the database, and store the data outside of salesforce.
What I need is to find the object/table that this field is stored in so I can write an SOQL query to get the field's values. Any ideas?
Easiest way is with Workbench.
I'm exploring all the options to persist user settings. The artilce in the url User-specific settings files for a windows form application: local xml file or database convinced me to store the settings into a db as my application is a standalone communicating with a DB. Is there any provider class like RegistrySettingsProvider to persist the data into database.
No - there isn't anything directly comparable. I'd look at the user settings as just another kind of data to store in the database. Use the same methods to store the user settings as you do the rest of the application's data.
I recommend NHibernate for your data layer. Just set up a mapping file and the database table and let NHibernate handle persisting your data to the database. Use that for all your database storage throughout the app and see how little database code you need to write.
I afraid no because your use setting is very application specific and hence there is no general to do the persistence.
You might want to consider one of the following two approaches:
Serialize your user setting class as a string, and store that string as a blob text column in one of your table.
Create a table schema that maps to your user setting, and persist the setting according to each column.
I would prefer the first approach because of its flexibility.