Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 9 years ago.
Improve this question
I want create a task as scheduler to get Data From SQL Server 2000 to Oracle ?
how i can do it ? thanks
This kind of processes are called ETL automation.
You should first write appropriate code accomplish the data transformation and extraction prior loading to other system. You have several options. One of them is using SQL. You may use either oracle to get data from SQL server or vice versa.
It is better to convert this code to a batch executable to able to run from command line.
Choose a scheduler to execute the job according to your business requirements.
Alternatives are cron for ux platforms, windows task scheduler for ms platforms, open source frameworks to implement your own (like quartz), open source products if you do not have budget or more professional commercial ones like TlosLite.
In Sql server you can create DTS job to transfer the data from SQL server table to Oracle table.Later using SQL server Agent,you can schedule this job.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I have few tables that I have to sync between 3 different RDBMS systems (PostgreSQL databases, a SQL Server and a Firebird Database).
Currently I simply connect to my Firebird database and pull the few relevant tables to my PostgreSQL database, but as databases change, new tables require querying and with the addition of a SQL Server database to the mix I feel this solution is ill fitting.
I've done some research on BI tools , but I still need to query data from this data source and show them inside a Windows Forms application.
PS: it's not a migration and I only need to query the data from these "satellite" databases
Using PostgreSQL as your hub, you can use Foreign Data Wrappers to reach out to the other two databases whenever a query wants their data. Then it will always be up to date, but performance might suffer compared to actually importing the data. For reaching SQL Server, you can use tds_fdw, and for firebird you can use firebird_fdw. I have never used either one of these, so this is just a starting point.
You could probably pick SQL Server as your hub and accomplish the same thing, it calls them "linked servers" rather than Foreign Data Wrappers, see for example.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I am pretty new to SSIS and SQL in general, and I am still learning the ropes. I am trying to create a few integration packages. First step would be to retrieve the data from one database and load it into another. I've been using the data flow task for this ... However the next step would be to run this package once a daily at a given X time (eg. 1 am). The landing database would be continuously updated , just like the source.
On a seperate package, I'm looking for a way to copy the data from my landing database, and dump it all into another table. This needs to be done so that the package automatically retrieves the rows from the landing database, before it gets overwritten everyday.
Any suggestions?
Thanks!
Regarding the scheduling of both the created SSIS packages they can be scheduled as SQL Agent job. SQL Agent is part of all SQL Server Editions except Express Edition. For details how to create the automated job to load SSIS packages, please refer to Schedule SSIS
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 6 years ago.
Improve this question
What necessary tuning I can do in ssis level,if all of the functionality have been
implemented using the Execute SQL task stored procedures.
All most all tables being used are already indexed.Please suggest some useful
ssis tuning tips for such kind of packages.Do I have to implement all these
execute sql tasks to Data flow task!?
To answer your question, no. Data flow tasks, in my opinion, are mostly useful when moving data between two sources or to do complex logic that you would rather use a c# script and not use the .Net CLR in SQL Server to perform that script.
If the source and destination are on the same instance I have never seen an occasion where SSIS data flow tasks can out-perform a properly built stored procedure.
Because you are using execute sql tasks, there is no SSIS performance tuning. All that SSIS is doing is launching something to be done by SQL Server.
I would focus on your stored procedures, look into the execution plans and read up on SQL Server Query Performance Tuning. It could range from a very simple index recommended when you look at the execution plan, moderate effort using Query Analyzer tool or very complex. It depends on your data and queries.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I have a SQL Server database with lots of data and an empty MySQL database. Now I want to copy all data of SQL Server database (including the database schema) to the MySQL database.
I thought about using raw SQL dump, but SQL Server's SQL syntax is not same as MySQL.
And I have googled some database migration tool, such as south (only for django), simple-db-migration (only for one type of database, maybe PostreSQL?) and SQLAlchemy (I haven't finished learning it, just feeling it somewhat clunky).
What tool can I use to do this migration?
I'm usually using opendbcopy for these kind of jobs ...
I tried this tool once http://dbconvert.com/convert-mssql-to-mysql-pro.php , but you have to pay for the full version. Of course you can write your own migration script e.g. in PHP. On Windows it will be easy to connect to MSSQL server, from a Linux machine you have to use ODBC and that is sometimes a bit messy to install and configure.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I'm using the SQL server 2008.
The scenario is:
We have about two hundred reports to generate daily, and it need plenty of time to query in DB and refine the data. We hope the reports can be sent from e-mail and saved at the windows file system at the same time. So the method need two different subscription for the two ways of delivery is not acceptable.
Wondering if we can make it by customizing the subscription directly or write some dll to enhance it(like allow us selecting a new customized delivery when edit the subscription). Could u pls give me some ideas on it? If not, we have to write a new program to do it..but it needs effort..
One option would be to cache the reports so that they run quickly the second time.
Another option would be to put the files in a Windows file share and use SSIS SendMail to pick up the files and email them out.
One more option is to write your own report runner that runs the reports and deals with them as you want. I was unsatified with the Reporting Services scheduler options and did it this way. It isn't as hard as you might think and Microsoft has some code that will get you started.