Repeated Header when exporting Dataprep recipe as CSV - csv

I am using Dataprep within Google Cloud Console. I am trying to export my recipe as a CSV. Export settings are the following:
The issue I am facing is that the final result is showing a replicate header, like the following, without any clear reason since the header should be present just once:
Any idea why this is happening? Any help would be much appreciated :-)
Marco

It's actually a bug. Trifacta is already on it:
Slack conversation:

Related

is there any way to automatically export reports to testrail?

I would like to export test data from RP to testrail automatically.
I have found the following issue but without any information on how it was solved or why was this closed. Would be happy to any guidance on how to set it up.
https://github.com/reportportal/reportportal/issues/145

Way to get detailed listing of all Pages properties in OneNote?

I have the latest OneNote on Windows 10. I'm trying to create a list of all my Pages in all the Notebooks. One line per Page, showing it's Name, Section and Notebook plus created date. As it's for import into (say) excel, be nice to have it delimited, say with # chars.
Nice Wines#Entertainment#Home Notebook#2021-10-01T06:57:32+10:00
I really only know Python and PowerShell. Is there any sample code you can suggest or another way to get this? Thanks.
I installed this PowerShell module and then another Stack Overflow user helped me get over another hurdle when using it.

No Google BigQuery table created after importing data through webclient

I'm currently familiarizing myself with Google BigQuery by working through the examples at https://cloud.google.com/bigquery/web-ui-quickstart. Doing a query over the pubic datasets runs fine.
I run into problems when uploading custom data into a new table through the WebUI. I create a new dataset and table, and upload the csv file provided with the example case. As in the example I input the schema and submit the file. Now the upload window stays on top and turns grey as if it's working. Nothing seems to happen afterwards though. When clicking away the upload window after a long wait, the table seems to be created in the tree on the left. However, when clicking on the table an error is shown:
"Unable to find table: ndwtest-984:csvtest.csvdata"
This seems like a trivial action, however I cannot seems to get it to work. I've tried varies different files, uploaded the file to Google Cloud Storage first and played around with the advanced options the last two days, but keep getting the same error.
Help would be much appreciated.
Some steps to help you:
billing must be enabled
you need to choose to upload one single TXT file from the example eg: yob2013.txt and not the zip file
make sure the schema is entered as text: name:string,gender:string,count:integer
on the last screen of the wizard you don't need to change the default CSV option parameters (for demo purposes works as it is)
I just tried the example, and it does work for me. In case you still have errors, than you can check your Job History menu in the Web UI, direct link would be, warning you need to put your Id in the link.
https://bigquery.cloud.google.com/jobs/YOUR_ANONYMOUS_PROJECT_ID_HERE?pli=1

Wiki special:export export 'Special pages' such as title=Special:DoubleRedirects

Is this possible or am I missing something? What I am attempting to do is quickly export list to xml or get into a csv. Attempted a myriad of strings and referred to the MediaWiki_Administrator's_Handbook to no avail. Any help or pointing me in the right direction would be appreciated.
You can get most of these queries through the api: see https://www.mediawiki.org/wiki/API:Main_page

BiarEnginer.jar/Command Line Import Documentation/Usage

Anyone know where the documentation is for the properites file?
I am trying the following exportQuery:
select * form ci_infoobjects where si_kind like 'FavoritesFolder' and si_name like 't%'
It is correctly grabbing the users but, is not biaring any of the folders/reports that said user has.
Suggestions?
Ok here is what I have found out.
documentation: on the command line importe is in the bi_vip pdf provided by SAP (which I will provide a link to later).
usage: My query isn't exactly correct but, removing the si_name portion should get you th right result.
issues: It seems that the command line importer does not gracefully handler export errors. Any export error it receives the command line will kill itself.
work around: I am pretty much duplicating the ImportWizard in java at this using API BIAROutput and IException.
I will hopefully post back my source code for the command line import tool in java.