kettle

How to deploy scheduled Kettle jobs on Pentaho BI server v6 CE

…衆ロ難τιáo~ 提交于 2020-01-02 14:40:02
问题 I have a server running Pentaho BI server v6 Community Edition. We've developed a Kettle job to extract from one database to another, exported as a KJB file. I would like to run this job every 12 or so hours. I noticed that the BI server already included Kettle, and has the ability to upload and schedule jobs. Do I need to install the DI server if the BI server already has Kettle installed? If not, how can I publish the KJB file into the BI server? I'd like to use a file system repository. If

How to implement database connection parameters validation in Pentaho kettle

核能气质少年 提交于 2020-01-01 16:47:20
问题 I have a case where I get X rows of database names as parameter which I need to connect to. I've succesfully made the job and transformations to repeat a table input step for each database name given as parameter. So this all works nicely when those dbnames happen to be valid. But if one of those names is invalid (db does not exist with given name), then the table input step fails with an error and the whole job stops instead of continuing for the rest of the names that are valid. So I'm

kettle表数据比较

你。 提交于 2019-12-26 23:16:20
使用合并记录组件://我的kettle死活不能保存中文,唉 其中tab_in_1和tab_in_2代表两个数据源, 合并记录: 新旧数据源可随意指定,获取需要对比的字段 此处为了对比将比较记录先放在file中 identical:比较的所有字段相同 deleted: 旧字段数据(旧数据源表有,新数据源表没有) new: 新字段数据(旧数据源表没有,新数据源表有) 也可以再加一个switch/case组件,将对应的记录分开存放: 通过指定flagfield字段,设置有相同记录标志identical的记录放入指定的文件。 来源: CSDN 作者: weixin_44135544 链接: https://blog.csdn.net/weixin_44135544/article/details/103722019

ETL JAR execution from BI server

感情迁移 提交于 2019-12-25 16:54:07
问题 I am trying to Execute a jar file from ETL - This works fine . When i am trying to call same ETL from xaction - This is showing errors as , ERROR 05-02 09:58:28,491 - Call Data Importer - org.pentaho.di.core.exception.KettleValueException: Javascript error:TypeError: Cannot call property runImageImpoter in object [JavaPackage com.MyTest.Data.Importer]. It is not a function, it is "object". (script#5) at org.pentaho.di.trans.steps.scriptvalues_mod.ScriptValuesMod.addValues(ScriptValuesMod.java

How to extract XML node values and from recurring nodes in pentaho?

半世苍凉 提交于 2019-12-25 03:13:57
问题 I am having a XML in the below mentioned format. I am using Get XML in pentaho to parse this. But I am not able get correct path to get all the values. I want all the values of ' custom-attributes '. It is only giving me the value of only first attribute .i.e for 'sex' only. Could anyone please help me how to get the values of other custom-attributes . I have given the ' LoopXpath ' setting as :- /names/name After pressing ' Get fields ' in fields tab as. I am automatically getting the custom

Pentaho Kettle - Get the file names dynamically

感情迁移 提交于 2019-12-25 01:47:23
问题 I hope this message finds everyone well! I'm stucked on a situation on Pentaho PDI Tool and I'm looking for an answer (or at least a light in the end of the cave) to solve it! I have to import, every month, a bunch of xls's files of differents clients. Every file has a different name (witch is given aleatory) and this files are on a folder named with the name of the client. However, I use the same process for all clients and situations. Is there a way to pass the name of the directory as a

How to validate one csv data compare with another csv file using Pentaho?

六月ゝ 毕业季﹏ 提交于 2019-12-24 20:11:13
问题 I have two csv file . In one file i have 10 rows and in another list of data . What i want to do is , check the data of one filed of first csv and compare it with another csv file . So how can i achieve this ? Any help would be great . 回答1: The step you are looking for is named the a Stream Lookup step.` Read you CSV and the reference files, and drop the two flows in a Stream Lookup and set it up as follow: a) Lookup step = the step that reads the reference b) Keys / field = the name of field

Pentaho Spoon Text File Output Additional Information Header

旧城冷巷雨未停 提交于 2019-12-24 15:31:05
问题 I am using the Text File Output step to create a CSV file, however i need to insert some additional rows of information at the top of the file. I have been able to have another transform output this data in a previous job step, however when doing so prevents me from outputting column headers in the appended csv output. The end result I am looking for would look something like this: EXTRACT TYPE: XYZ DATE: 20110520 FIRST NAME,LAST NAME,AMOUNT charlie, chaplain, 2345 someone, else, 1234 Any

kettle:一个数据库错误发生在从资源库文件读取转换时

佐手、 提交于 2019-12-24 15:20:39
问题情况:   现场反应导入一个资源库脚本后,打开导入的转换再保存后,再次打开报错:一个数据库错误发生在从资源库文件读取转换时   先贴日志吧(手输的,将就了): nullpointexception: at org.pentaho.di.repository.kdr.delegates.KettleDatabaseRepositoryStepDelegate.loadStepAttributesMap(KettleDatabaseRepositoryStepDelegate.java:441)   先打开代码看看,发现是步骤属性读取为空。 private Map<String, Map<String, String>> loadStepAttributesMap( ObjectId stepId ) throws KettleException { Map<String, Map<String, String>> attributesMap = new HashMap<String, Map<String, String>>(); List<Object[]> attributeRows = repository.connectionDelegate.getStepAttributesBuffer(); RowMetaInterface rowMeta = repository

Database connection to Neo4j with Pentaho Spoon

放肆的年华 提交于 2019-12-24 15:11:29
问题 I want to connect to Neo4j with Pentaho Kettle Spoon. I downloaded jdbc driver of neo4j with this and I tried to use this guide to connect to neo4j with Pentaho Kettle Spoon. I have two main problems: downloaded JDBC Driver is zip file. I changed Driver extension to jar for solving this problem. when I changed extension to jar and copy it in lib folder and follow this guide I faced missing driver error: org.neo4j.jdbc.Driver could not be found How can I solve these problems? 回答1: You can not