etl

Easiest way to extract SharePoint list data to a separate SQL Server table?

这一生的挚爱 提交于 2019-11-30 15:08:09
问题 Edited: What is the easiest way to scrape extract SharePoint list data to a separate SQL Server table? One condition: you're in a work environment where you don't control the SQL Server behind the SharePoint Server, so you can't just pull from the UserData table. Is there there any utilities that you can use to schedule a nightly extract? Is Microsoft planning any improvement here for "SharePoint 4"? Update Jan 06, 2009: http://connectionstrings.com/sharepoint For servers where office is not

Easiest way to extract SharePoint list data to a separate SQL Server table?

家住魔仙堡 提交于 2019-11-30 13:46:50
Edited: What is the easiest way to scrape extract SharePoint list data to a separate SQL Server table? One condition: you're in a work environment where you don't control the SQL Server behind the SharePoint Server, so you can't just pull from the UserData table. Is there there any utilities that you can use to schedule a nightly extract? Is Microsoft planning any improvement here for "SharePoint 4"? Update Jan 06, 2009: http://connectionstrings.com/sharepoint For servers where office is not installed you will need: this download As Ryan said I would also suggest using object model / web

How to add third party libraries to Talend project?

前提是你 提交于 2019-11-30 12:11:04
How to add third party libraries (jar files) to a Talend project ? One more question is, Each Talend component uses LogFactory , but in my case it is throwing java.lang.NoClassDefFoundError: org.apache.commons.logging.LogFactory at ... with tSetGlobalVar component, how can I make use of existing logging library to get rid of this exception? Vikram Takkar You can import the external jars using tLibraryLoad component or Editing Routine Libraries . For more information visit http://vikramtakkar.blogspot.in/2013/03/importing-external-jar-or-library-files.html Just to benefit others, with recent

orientdb load graph csv of nodes and edges

流过昼夜 提交于 2019-11-30 09:57:11
问题 I'm a newbie in Orientdb . I have a csv file which has both the nodes and the edge and I need to create a graph out of that csv file . csv file "p1","p2","score" "LGG_00001","LGG_01682",282 "LGG_00001",".LGG_01831",183 "LGG_00001","LGG_01491",238 The edge is IsActingWith which had the score attribute { "source": { "file": { "path": "C:/Users/sams/Desktop/OrientDB2/lac2.csv" } }, "extractor": { "csv": {} }, "transformers": [ { "vertex": { "class": "lac2" } }, { "vertex": { "class": "lac2" } },

Add incremental number while inserting records using SSIS

为君一笑 提交于 2019-11-30 09:48:36
问题 I have a SSIS package in which, two records are coming. I need to insert the records in the table with an extra column (let's say Sequence). If there are two records, Sequence column should have the value 1(for the first record) and 2(for the second record). Again, next time, I'm getting three records, then again sequence starts from 1,2 and 3. Is there anyway to do this without using script or stored procedure? Screenshot: 回答1: There are 2 methods to achieve this: (1) why not using a script

Replace missing values in SAS

让人想犯罪 __ 提交于 2019-11-30 08:37:39
问题 How do you replace all missing values with zeroes in SAS? I have a text file that I dump into SAS to process some geo data, but whenever it has a missing value it breaks the operations. Is there a way to change this without specifying each field? I have over 200. The way I do so is: data geo_cali_north; set geo_cali_north; if polar_data eq . then 0; if lat_xvar eq . then 0; run; How can I avoid doing this for every field? 回答1: You can set all the missing values to 0 with like this: data

What is a staging table?

感情迁移 提交于 2019-11-30 08:17:17
Are staging tables used only in Data warehouse project or in any SSIS Project? I would like to know what is a staging table? Can anyone give me some examples on how to use it and in what circumstances it is implemented? Also, may I please know the best practices while using it? staging tables are just database tables containing your business data in some form or other. Staging is the process of preparing your business data, usually taken from some business application. For your average BI system you have to prepare the data before loading it. A staging table is essentially just a temporary

SSIS Data Flow Task hangs on excecution of Pre-excecute phase

左心房为你撑大大i 提交于 2019-11-30 07:59:44
I have a Data Flow Task that is hanging on excecution. The flow is simple, makes two queries to different tables (Both with a couple of joins), then sorts and merges the otuputs through a common id, adds a static column to all the records, saves the row count in a user variable for later use and finally inserts into a table on another DB. We are using OLE DB Sources and Destination. Source is MSSQL 2000 and Destination is MSSQL 2012 Symptoms: When excecuting, the Data Flow gets the usual yellow "running" icon. However when you double click to see the Data Flow, non of the elements have any

数据迁移实战:基于Kettle的Mysql到DB2的数据迁移

家住魔仙堡 提交于 2019-11-30 00:58:11
一、什么是 ETL ETL,是英文 Extract-Transform-Load 的缩写,用来描述将数据从来源端经过抽取( extract )、转换( transform )、加载( load )至目的端的过程。 ETL 一词较常用在数据仓库,但其对象并不限于数据仓库。 二、Kettle简单说明 Kettle是一款国外开源的 ETL 工具,纯 java 编写,可以在 Window 、 Linux 、 Unix 上运行,数据抽取高效稳定。 三、需求 1、 Mysql 5.X数据迁移到 DB2 8.X 2、 主键自增改成 32 位 UUID 三、准备工作 1、 源数据库: Mysql 5.6 2、 目标数据库: DB2 8.X 3、 数据迁移工具: Kettle6.0.0.0-353 四、实战开始 1、说明一下省略掉的步骤 在开始使用 Kettle 前,我使用了 PowerDesigner18.6 建立数据模型,然后导出对应数据库的建表脚本,这里我导出了 DB2 8.X 的建表脚本,然后首先在 DB2 中将目标表建立好。其实在 Kettle 中也可以在执行过程中来建立表,我也是初学 Kettle ,所以选择了一种相对简单的方式来操作,毕竟关注重点在于数据迁移这方面上。 2、安装 JDK 因为 Kettle 是纯 JAVA 写的,所以依赖 JDK ,至于 JDK 的安装不是本文的重点

ETL SSIS : Redirecting error rows to a seperate table

99封情书 提交于 2019-11-29 22:07:39
问题 I am working on a package that contains a Source, about 80 lookups and 1 destination. The data in the source table is not consistent enough and hence my package fails very often. Is there a way by which I can transfer all the rows which are giving at the time of inserting them in destination table? For eg. I have 5 rows in Source and out of which 1st and 4th will give error. Now the result should be that 2nd, 3rd and 5th should go in destination but 1st and 4th should be stored in some flat