Questions tagged [pentaho-data-integration]

Tag to be used for Pentaho Data Integration (all versions). Pentaho Data Integration prepares and blends data to create a complete picture of your business that drives actionable insights.

Pentaho Data Integration prepares and blends data to create a complete picture of your business that drives actionable insights.

It provides intuitive drag-and-drop data integration coupled with data agnostic connectivity spanning from flat files and RDBMS to Hadoop and beyond.

Features:

  • Graphical extract-transform-load (ETL) designer to simplify the creation of data pipelines
  • Rich library of pre-built components to access, prepare, and blend data from relational sources, big data stores, enterprise applications, and more
  • Powerful orchestration capabilities to coordinate and combine transformations, including notifications and alerts
  • Agile views for modeling and visualizing data on the fly during the data preparation process
  • Integrated enterprise scheduler for coordinating workflows and debugger for testing and tuning job execution
825 questions
0
votes
0 answers

How to access pagination values in kettle from pentaho cde?

I have a table component with data source attached to kettle transformation. Now, I want to implement server side pagination and access the parameters "pageStart" and "pageSize" set by pentaho in kettle transformation using getVariables. I have…
0
votes
1 answer

How to set content-type in Rest Client step

I'm using the Rest Client as below, But every time I have the same error message HTTP Error 400. The request has an invalid header name. When I use the same config in postman client, (Content-Type:application/x-www-form-urlencoded) I've success…
luiz kava
  • 187
  • 1
  • 2
  • 14
0
votes
2 answers

Applying Left Join in Pentaho

I'm try to create Transformation and need to merge two Database based on query like that by using Merge Join and I little bit confuse what should i filled in First Step, Second Step to Lookup for that each query format. Query Format : SELECT * FROM…
Rio Odestila
  • 125
  • 2
  • 19
0
votes
0 answers

Migration from Pentaho Data Integration to Talend Data integration / Big Data

I wanted to know the feasibility of migration of group of data processes on Pentaho Data Integration to Talend. By knowing databases are in POSTGRESQL. It is possible ? Complicated ? Impossible ? If it is possible, what are the requirements ? My…
0
votes
3 answers

Error in Pentaho Data Integrator - invalid byte sequence for encoding "UTF8": 0x00

Error getting while insert bulk rows with Pentaho Data Interrogator. I am using PostgreSQL ERROR: invalid byte sequence for encoding "UTF8": 0x00
Md Sirajus Salayhin
  • 4,974
  • 5
  • 37
  • 46
0
votes
1 answer

PDI - PosgreSQL Bulk Load error, Insert/Update and Table Output slow

i wanna ask about this : BTW, I'm new with PostgreSQL and PDI. I have search about this in internet and I don't get it how PostgresSQL Bulk Load work and why this error (image below) show up. PostgreSQL Bulk Load Error This error show up when I try…
Rio Odestila
  • 125
  • 2
  • 19
0
votes
1 answer

Error Showing While Running Pentaho Data Integrator Community Edition on OSX

While running spoon.sh from my Mac OS I found following error: After a long time spoon open but there is no option in Design or Transformation. I am using OSX 10.13.5
Md Sirajus Salayhin
  • 4,974
  • 5
  • 37
  • 46
0
votes
1 answer

Pentaho add Last Modification Field in Output Table

In Pentaho input step, there are the tab Additional Output Field. I want to use Last Modification Field in output. What should I write in Last Modification Field so in output it will write timestamp. (Please give example also) Thank you.
Haha TTpro
  • 5,137
  • 6
  • 45
  • 71
0
votes
0 answers

what value type to be used to check/unchecked a checkbox in PDI ETL Metadata injection

I have ETL Metadata Injection steps which injects the meta data to the CSV file input step present in the template. I am able to inject all the meta data like File name to read , delimiter and field names and type of fields but I am not able to…
Ghost Rider
  • 688
  • 3
  • 17
  • 38
0
votes
2 answers

PDI - GC Overhead Limit Exceeded

So, I have some questions about PDI and my Transformation flow. Based on my trans flow I get error GC Overhead Limit Exceeded and I have already search for solution like increase the memory for spoon.bat -Xms -Xmx2g but it doesn't work. So I think…
Rio Odestila
  • 125
  • 2
  • 19
0
votes
1 answer

PDI 7 o 8 error connecting to SQL Server repository

I have a SQL Server database repository. There are a lot of Jobs and Transformations on it. If I use PDI 4.4.0, I can connect and to work on it, normally. I can create or edit any transformation or job, no problem. But, when I tried to use PDI 7.1…
0
votes
2 answers

Pentaho: combination lookup/update does not process all rows in the source

I am using Pentaho Data Integration to do a SCD type 1 transformation. I am using combination lookup/update transform to generate the surrogate key value (upon insert). The commit size is 100000 and the cache size is 99999. My source table has 19763…
Kuldip.Das
  • 191
  • 1
  • 3
  • 13
0
votes
1 answer

Pentaho start a java program inside the same jar

We have an organization reports portal for which we use Pentaho. Now the catch is, we need to define one single java Jar in which Pentaho + java files + sql, all are there. I'm able to execute SQL via Pentaho but I need to execute a class in the…
0
votes
1 answer

Pentaho-kettle to write log in a file using java

I have a kettle job and it's running successfully using java. But, kettle writing it's log into standard output (console). I want logs to be written in any file (.txt, .log). Is there any way to write logs in a file? And if possible with the user…
0
votes
2 answers

How to start inserting row after some specified number to MySql database using Pentaho?

Basically what i want to do is that, I have CSV file containing 10,000 rows that i want to insert into the database . When i start my transformation i want to start inserting in database after 4500 rows . So i want to skill number of rows that i…