pentaho

Kettle: Multiple putRows() in processRow() correctly?

末鹿安然 提交于 2019-12-09 21:15:02
问题 I'm processing a /etc/group file from a system. I load it with CSV input step with the delimiter : . It has four fields: group , pwfield , gid , members . The members field is a comma separated list with account names of unspecified count from 0 to infinite. I would like to produce a list of records with three fields: group , gid , account . In the first step I use User Defined Java Class , in the second I use Select values . Example Input: root:x:0: first:x:100:joe,jane,zorro second:x:101

JDBC converting Timestamp to NULL (zeroDateTimeBehavior issue)

拈花ヽ惹草 提交于 2019-12-09 18:01:08
问题 I'm using Pentaho Data Integration (Table Input step) to pull in data from a MySQL server. A couple of fields are of the type 'Timestamp', and Pentaho keeps spewing out errors because of the timestamp being NULL (0000-00-00 00:00:00.000000). I added a zeroDateTimeBehavior=convertToNull to the parameters which should take care of the bad timestamps, but it's converting all of my Timestamp data to NULL. One reason why I think it may be happening is because some of my 'good' data is represented

Pentaho CDE “All” option in “Select” component

社会主义新天地 提交于 2019-12-08 10:14:09
问题 While a "Select All" feature in a select dropdown component would seem to be a straightforward feature, I've had an incredibly hard time finding a workable solution. I can find three SO questions on the topic, but none of the approaches work entirely. How to add a select all option to select component in CDE Pentaho CDE reset parameter How to add a "select all" option to select component in CDE? I have a table with a column that contains ratings from 1-5 (as string-type). I need to give users

integration Pentaho BA 7 community edition with CAS

时光总嘲笑我的痴心妄想 提交于 2019-12-08 06:47:55
问题 Recently I am working to integrate Pentaho BA (BI server) 7 CE with Cas server . I follow up Pentaho official documentation . https://help.pentaho.com/Documentation/5.1/0P0/150/010/060/000 but unfortunately it doesn't work!! Actually it's kind of hard to find about section 5 in documentation . where I have to set casAuthenticationProvider.MemoryUserDetailsService in my configuration ? And all i've get from the engine is the error which says it couldn't found CasAuthenticationProvider class .

Create Datasource Hive on Pentaho hadoop

烈酒焚心 提交于 2019-12-08 06:35:58
问题 I try pentaho hadoop to make report designer with Hive Datasource. I have make connection to hadoop with connection url: jdbc:hive://localhost:10000 and driver class name :org.apache.hadoop.hive.jdbc.HiveDriver Then I create datasource which I pointed to Hive database to access Table inside. And then when I make design the report and drag field of table to designer workspace there is and error message An unexpected error has occurred: The generated SQL-query did not execute successfully. in

In pentaho..How to pass a text file which contains all the definition of the connection parameters in the job?

偶尔善良 提交于 2019-12-08 06:22:39
问题 I am using jdbc connection and i am passing parameters with example ${sample_db_connection} and that parameters has been defined in server in a text file as sample_db_connection=localhost and i want to pass the text file in the job step so that whenever the job ran and it found this parameter ,automatically it will take the value defined in text file. 回答1: You need to create a KTR file using " Property Input " as the input step and " Modified Java Script " Step to define the key value mapping

Pentaho HTTP Post using JSON

早过忘川 提交于 2019-12-08 03:40:23
问题 I'm brand new to Pentaho and I'm trying to do the following workflow: read a bunch of lines out of a DB do some transformations POST them to a REST web service in JSON I've got the first two figured out using an input step and the Json Output step. However I have two problems doing the final step: 1) I can't get the JSON formatted how I want. It insists on doing {""=[{...}]} when I just want {...} . This isn't a big deal - I can work around this since I have control over the web service and I

Kettle '?' not working Table Input Step

喜夏-厌秋 提交于 2019-12-07 15:20:46
问题 I want to get all the table names from the database and then get all the rows from the tables. So I created a transformation like this: Get Table Names: Added the database connection and stored the table name in a output field called "tablename". Table Input: Marked "Replace variables in script" and "Execute for each row". Added the first step in "Insert data from step". SQL is "SELECT * from ?". I have read up a lot of tutorials online, including the documentation. My problem is that

Using PDI transformation in Pentaho BI server as data source for report parameters

人走茶凉 提交于 2019-12-07 14:32:49
问题 Any advice on how to use PDI transformation as data source for report parameters in BI server's console? I've uploaded the prpt reports to BI server but the I get a message "Error parsing parameter information". The .prpt and .ktr files are both in the same directory. 回答1: Actually, just realized that the issue could be solved by adding Transformation (KTR) as a resource. In this case, one can use the File-Resources menu selection. In the dialog select the transformation you wish to import

Filling data gaps in a stream in Pentaho Data Integration, is it possible?

血红的双手。 提交于 2019-12-07 07:13:09
问题 I have a CSV file with currency exchanges EUR-USD. The file was downloaded from the Bank of Canada. I downloaded the CSV with data since Oct 10th, 2013 onwards. There are, nevertheless, gaps in the data, ie. days without the conversion rates. I've been fighting (1st day with Spoon Kettle) to find out a simple (but general) way to fill the gaps, say, with the last non-null value. And the only way I've managed to accomplish this is by chaining 4 "Get previous row fields" and the using the NVL