SAP Gateway projects: How to re-anable import functionality of Date Models after project copy? - sap-gateway

Using t-code SEGW, I created a SAP Gateway project with several data models based on RFC function modules. The data structure is based on the handy "Import" funtionality.
Due to a customer request, I had to copy the project into another, customer owned, namespace.
After the copy, I lost the funtionality to import fields to the data models.
When I create a new entity, the possibility to import is still there. When I copy a fully working project, lets say to $tmp, I lose the import feature again.
Sure, I can still edit my entities by hand but it is tiresome and prone to errors.
So, my question, is there a way to re-enable the import feature of copied SAP Gateway projects?

Check if you are importing complete project into customer system, check for the service and model. It should have come if the project has been imported successfully.
Note: Please update SEGW with latest release.

Related

Create and make analysis CortexDB

I expose my problem: I have recently started using CortexDB, a NoSQL software to database analysis. I have read the (poor) documentation on https://docs.cortex-ag.com/en/CortexDB/CortexDB/, and purchased a free license to evaluate the operation of the program. As the documentation is unclear I would have some questions to ask you:
1) How do I create a database?
2) how can I import a database contained in an excel file (.csv)?
3) how do I create charts or analyzes regarding the data entered?
Thanks
because the question is very old I hope I can still help you.
First of all: you should download the latest release of the free version (simple registration and download)
if you downloaded the free version you got the server and two databases. A server process handles one database. For a second database you have to start a second server (different port of course). If you start the free version you should have an empty database (or the filled and configured demo db). If you want to create a complete new one without any predefined configuration, you have to start the server process with the command line and the parameter -n (ctxserver64 -n). If you did that you have to configure everything by hand with the tool ‘remote admin’.
the question is not clear for me. Do you mean how to import a csv file into a CortexDB or do you mean how to import the database content into an excel file?
If you want to import the csv file into a CortxDB, the easiest way is to use the tool CortexImplex. It’s completely explained in the online docs (https://docs.cortex-ag.com/en/CortexImplex/CortexImplex-Basics/)
If you want to export datasets as csv file the only thing you have to do is to configure a list in the CortexUniplex as a view for your datasets and export them as csv (you find the export function in the list menu).
I would do the charting with d3j. For this you can use the so called ‘DataService’ of the CortexUniplex. It’s a kind of an API for posting requests and getting JSON objects. If you have a completely configured UniPlex you can use all of your configuration as json objects for other apps (for example charts or an individual application).
The full version has a simple dashboard inside of the CortexUniplex. Maybe the vendor offers it in the free free version.
By the way: it’s always good to write an email to the info address. Because this database is not so famous and known, the guys are very helpful. Or contact them via twitter or other channels (see at the bottom of the cort ex-ag.com webpage).

Is possible to connect the company DB to Odoo 9 (OpenERP)?

I am creating a business app to manage the inventory of a mid-sized shoe company.
I already have a database created with PostgreSQL. I would now be able to connect it to Odoo 9 to develop applications.
How can I do that?
You can't just use your existing data directly with odoo
But Odoo makes it easy to use existing data from existing applications and databases with it's flexible import system. in which data is imported to it via csv files, you can import any data you can possibly think of, you just have to know the right format of data that odoo is expecting and the required fields you have to provide
Every model in Odoo even custom built modules can have data imported into them
If you want to import data just click on the listview of any model and you should see an import link beside the create button, from there you can carry on, there's even a FAQ there which you can read if you need to find out more about the import system.
As for knowing the required columns that have to be present when you're importing, Just create sample data and save it, Then
Go to the listview
Tick the check-box beside the record name to select it
Now click on the more button you should see an export option
From there you can pick the fields you want to export, the required fields will highlighted in blue
There are other ways of re-using existing data with odoo (You can write to the database directly...but this is not recommended as several constraints would not be checked and you can end up messing the database up) or you can use xmlrpc or jsonrpc which is going to be time consuming
Using the import functionality is the best and the easiest way.

Where are a node data stored in Drupal?

I have a working Drupal site that i need to replicate on another server, so that the replicated version doesn't contain custom site information like users data or specific article published on the main site.
So i did a selective dump of the main db, and re-imported in the replicated site. In the specific case of the "node" table, I imported only nodes of the type "page". Now, all went well. The main functions of the site are working, but i have a problem with pages throwing an error "The requested page "/node/page_number" could not be found." with ALL the basic node pages. I can see only views and the front page.
Probably i'm not dumping some information needed by Drupal to display the node. So the question is: where i can find info about basic nodes? I thought porting the table "node" and "node_access" was pretty enough, but really it isn't.
The problem is that node table stores only basic information.
I think you did not imported node_revision table as well.
Even if you do, dumping tables and re-importing only specific data into another database breaks their connections, by resetting id columns and missing extra data provided by fields and other custom modules.
Consider using Import/Export modules, comparison of which you can find Comparison of Content and User Import and Export Modules
Consider the data_export_import module.
What I like about data_export_import is that it is an all-in-one module (for export AND import of data. Whereas as with some other solutions, the import and export stages are done by separate modules, which could mean that extra work is needed ensuring compatibility of the export with the import.)

How can I import Landslide CRM data into Salesforce?

I'm importing some customer information from Landslide CRM into Salesforce.
Anyone have advice on the best methodology for doing the import?
It seems like the Apex Data Loader is the best way to go, but I don't know
if there are any issues with handling the objects in question, or if there
might be a specific tool or script to perform this migration.
Any experience with this import in specific or importing data into Salesforce
in general would be appriciated.
Importing Data to salesforce can be achieved in multiple ways depending on the type of data nd the requirements you have.
The first thing to do is get your data into CSV files so you'll need to find a way to export the dat afirst. For UTF-8 encoded data don't use Excel use something like OpenOffice (only required if you have UTF-8 Characters)
If its account and contact data for example. There is an import wizard available in Setup > Administration Setup > Import Business Accounts/Contacts
Next Option is as you say to use the Apex Data Loader. This probably the best approach.
The first thing and this is critical for big migrations is to Create a Field on your account object which will be a Unique Field for reference purposes. When creating this field set it as an External ID field and populate it with a unique reference for your accounts, the same goes for anything else which will be a parent. (you'll see why shortly.)
Next use the Insert option in the Data Loader to load the data mapping all the fields, especially the External Id
Now when you upload child objects use the Upsert option and map your Account Id via the External Id created earlier. This will match the accounts using your unique Id instead of you having to use the Salesforce id, saves alot of time.
Repeat the same for other objects and you should be good to go.
Apologies for the lack for structure here... doing this while in work and don't have alot of time but hope this helps.
The data loader works great for most types of imports. The one suggestion I would give you is to create a new custom field on your target objects (presumably Account and Contact) called "Landslide ID" or similar, identify it as an external ID field, and then import the primary keys from your source system into this field (along with the "real" data).
Doing this achieves a couple things - first, you have an easy unique link back to the source data for troubleshooting or tracing back to the source system. Second, if you find yourself in a situation where you need to import more fields or related data from the original source system, you'll be able to do so in an easy and correct way. It's just a good standard practice to adopt when doing data migrations -- it's almost no additional effort and can save you many hours in the future.

Best strategy to initially populate a Grails database backend

I'd like to know your approach/experiences when it's time to initially populate the Grails DB that will hold your app data. Assuming you have CSVs with data, is is "safer" to create a script (with whatever tool fits you) that:
1.-Generates the Bootstrap commands with the domain classes, run it in test or dev environment and then use the native db commands to export it to prod?
2.-Create the DB's insert script assuming GORM's version = 0 and incrementing manually the soon-to-be autogenerated IDs ?
My fear is that the second approach may lead to inconsistencies for hibernate will have the responsability for the IDs generation and there may be something else I'm missing.
Thanks in advance.
Take a look at this link. This allows you to run groovy scripts in the normal grails context giving you access to all grails features including GORM. I'm currently importing data from a legacy database and have found that writing a Groovy script using the Groovy SQL interface to pull out the data then putting that data in domain objects appears to be the easiest thing to do. Once you have the data imported you just use the commands specific to your database system to move that data to the production database.
Update:
Apparently the updated entry referenced from the blog entry I link to no longer exists. I was able to get this working using code at the following link which is also referenced in the comments.
http://pastie.org/180868
Finally it seems that the simplest solution is to consider that GORM as of the current release (1.2) uses a single sequence for all auto-generated ids. So considering this when creating whatever scripts you need (in the language of your preference) should suffice. I understand it's planned for 1.3 release that every table has its own sequence.

Resources