How tfileinputexcel works in talend - talend

I'm trying to load xlsx file into postgresql database.
my excel file contiens 11262 row but after excuting the job i fond 14**** rows and i don't know why? I want only to find my 11262 rows in my table
here is the job
here is my excel file

I'm guessing you didn't check Stop reading on encountering empty rows, and that it's reading empty rows from your excel file.
Please try with checkbox checked.

Related

Upload multiple files to pentaho

In pentaho data integration, how do I import a list of xlsx files that are in the same folder?
note: the number of columns are always the same
If your excel column name and sheet name are always same then you can use THIS solution. Here I have take all xlsx file from source folder and convert one-by-one file as csv.
But if your excel columnname and sheet name are dynamic or you need some dynamic solution. Then you can use my another stackoverflow solution from Here

How can we Exclude the unnecessary rows from Excel File while doing Data Load using Copy activity in ADF

I have a excel file which is semi-structured. There is data in a table, but there are dividers in certain rows that should needs to be ignored.
The processing of the data should start with the column headers(Col1 , col2 ....) and only process the rows with actual data.
Could anyone suggest the way to achieve this using copy activity in adf .
My source is xls file and target is ADLA (Parquet file)
Any help appreciated. Thanks in advance.
The most closest solution is that you need manually choose data range in the excel file:
Ref: https://learn.microsoft.com/en-us/azure/data-factory/format-excel#dataset-properties
HTH.

Dataprep : Invalid array type after run job to excel file

I try to use array type column in dataprep and it is look good in dataprep display ui as the picture below.
But when I run job output with .csv file, there are invalid value in the array column.
Why does the .csv output different from dataprep display?
Array in Dataprep display
Array in csv output
It looks like these two columns each contain the complete record...? I also see some non-English characters in there. I suspect something to do with line breaks and/or encoding.
What do you see if you open the CSV file in a plaintext editor, instead of Excel?
What edition of Dataprep are you using (click Help => About Dataprep => see the Edition heading)?
What version of Excel are you using to open the CSV file?
Assuming that this is a straight-forward flow with a single dataset and recipe, could you post a few rows of data and the recipe itself (which you can download), for testing purposes?

How can I extract 7 million records from Oracle database to CSV file

I have already tried the export option in the sql developer,it is very time consuming.
I need to know if there is quicker way to extract data to CSV file.
enter image description here
I think you can try with Toad for oracle in DataGrid view then right click and export Data Set, you can start with a million with WHERE ROWNUM < 1000001

Space in column name Talend

I want to make a csv file that I can upload in my Google Calendar.
The mandatory headers for a file to upload are
Subject, Start date, Start time
But in Talend you can't make a column name with a space between the words, anybody know how I can fix this?
Maybe you can generate the first line with a "tFixedFlowInput" and complete your CSV file without column titles by changing in your output component the parameter "Include Header".
Don't forget to check the parameter "append" when you insert your data after