I'm very new to Hadoop. I want to connect SQL Server 2008 to Hadoop.
Can anyone give me the steps to do this? Anyone answering this is highly appreciated.
You can use Apache Sqoop for this purpose.
you can use Sqoop for efficiently transferring bulk data between Apache Hadoop and structured data stores such as relational databases.
Use the following link to install Sqoop
Related
I'm unable to add or connect MS SQL server remote Database in Neo4j?
I've tried several times by adding the remote server URL in the Neo4j tool
Also, how do we make this tool available to everyone or end users without installing on their devices?
Can anyone please guide me?
Please treat this as a request.
https://community.neo4j.com/t/how-to-connect-a-remote-sql-server-in-neo4j-tool/37962?u=kdplghna
As mentioned on the community post, Neo4j Desktop is not a general db management app, it is specifically for managing and working with Neo4j, so MS SQL support will not be added. Please search for other tools that are designed to work with that db.
If you want to connect to an MS SQL server from Neo4j, then you could use APOC to perform a load via JDBC:
https://neo4j.com/labs/apoc/4.1/database-integration/load-jdbc/
I am trying to sqoop export data to SQL Server with --update-mode as 'allowinsert' and getting below error :
Error during export: Mixed update/insert is not supported against the target database yet
Has anyone faced this error with SQL server and is there any version of SQL Server which supports upsert operation through SQOOP. The Sqoop version I am using is 1.4.6
I'm facing the same issue. The feature is simply not supported in Sqoop 1.4.6 (for SQL Server that is). The best way to do so is, if possible, upgrade Sqoop to version 2.
In my case that is not possible, so I have to perform two sqoop jobs: one for inserting and one for updating.
If someone have a better way of doing that, please feel free to share!
It appears to be working with the following combination:
Sqoop 1.4.6.2.6.5.0-292 and SQL Server 13.0.5026.0 (SQL Server 2016). But I'm not sure that it's working because of the specific combination, rather than specific settings on the SQL server OR this specific version of Sqoop.
I am using WebSphere application server for deploying my EAR Application . I have DB2 Database in that server . Now i am trying to Deploy same Application on Jboss application server with PostgreSQL. I need to change the existing DB2 database to PostgreSQL . I have researched about this topics and finally i got some useful tools .
https://github.com/dalibo/db2topg
https://wiki.postgresql.org/wiki/Converting_from_other_Databases_to_PostgreSQL
Now i want SQL dump of my DB2 schema for converting DB2 to PostgreSQL . I want to know whether can i get SQL dumps of DB2 from the application server of WebSphere ? Or any other methods to convert DB2 to PostgreSQL ? Any suggestion will be helpful to me to solve this problem .
No, WebSphere Application Server does not provide a general means to export the schema of a DB2 database. A google search for export db2 schema will provide you multiple ways to do it using DB2. Again a google search for convert schema from db2 to postgres yields numerous tools to migrate schemas including a postgres wiki.
I have a requirement of pulling data from sybase database to SQL using an SSIS package. I cannot find any provider/sybase connector till now. Searched a lot on google/tried multiple suggestions but picture is not clear anywhere.
Kindly let me know what is the step by step process and download link of driver if required. I would prefer if it can be done without any new installation as it would require more approvals from client.
Install ASE from SAP website, then using ODBC connector, we can connect to sybase database in SSIS.
I chose to use Apache Mahout as my recommendation engine but at the same time due to some reasons it would be easier if I could store my data in a SQL Server db. Can mahout be connected with SQL Server without any problems ?
The documentation says that it can be connected with other db engines through JDB driver but I see all articles , books using mysql and also the data model supported are for mysql only.
How to convert MySQL to SQL Server databases:
SQL Import/Export Wizard through ODBC (http://www.mssqltips.com/sqlservertutorial/2205/mysql-to-sql-server-data-migration/)
SQL Server Migration Assistant (http://msdn.microsoft.com/en-us/library/hh313125(v=sql.110).aspx)
Here is the JDBC driver for SQL server:
JDBC Driver for SQL Server: http://msdn.microsoft.com/en-us/sqlserver/aa937724.aspx
Changing DB input format/driver in Hadoop cluster: http://blog.cloudera.com/blog/2009/03/database-access-with-hadoop/
There are also numerous example of using Mahout with an Azure Hadoop Cluster via HDInsight:
http://bluewatersql.wordpress.com/2013/04/12/installing-mahout-for-hdinsight-on-windows-server/
http://www.codeproject.com/Articles/620717/Building-A-Recommendation-Engine-Machine-Learning
I have just started my experiments with Mahout. I managed to run some book examples after replacing the in-memory data models with SQL92JDBCDataModel or SQL92BooleanPrefJDBCDataModel shipped with Mahout 0.9.
I passed an instance of SQLServerDataSource to constructors of those data models. This class is included into the Microsoft JDBC Drivers for SQL Server package (I used the version 4.1)
However, the SQL92JDBCDataModel documentaton states that it is "not optimized for performance".