Pentaho behaves slowly on insert/update with PostgreSQL, which is because jdbc driver will not accept prepared statements.
After 2 million row Pentaho / PostgreSQL performance almost become 1 per 1 second. Post me an alternative solution with example.
Below is my Create Table script
CREATE TABLE fact_model
(
model_id integer NOT NULL,
source integer NOT NULL,
group_sk integer NOT NULL,
user_sk integer NOT NULL,
order_kind character varying(255),
model_no character varying(255),
CONSTRAINT pf_fact_receipt PRIMARY KEY (source, model_id)
)
WITH (
OIDS=TRUE
)
ALTER TABLE fact_model
ADD CONSTRAINT pf_fact_receipt PRIMARY KEY(source, model_id);
Guys, it is crazy replies I am getting, My expected intention was to help other, There are various step we can try. like follows.
- Insert / Update which is slow
- Upsert after PostgreSQL after 9.5 using SQL step
- Update then error handled to a table output when we expect higher Update
- Insert then error handled to Update when we expect higher Insert
- Merge join split insert and update direct to appropriate step which makes one single
I believe there are always Lot more to figure out beyond my experience which I believe this post will sort out., So the expectation is better input than stated above.