CSV File creation affected by regionnal setting - csv

I create csv file from my web app and when the web app runs in my system with browser language set to English the file gets downloaded and on open shows the expected format. But when i change my language to Danish it shows like below !
Wrong csv file format
The correct format should be like this
How do i correct this so that irrespective of the language the language the file csv file gets downloded properly ?
Thanks in Advance

I'm not sure about the components you use to create, download, and display the .csv; so this may not help.
In general, language settings may specify the separator used for .csv files. If you can generate the .csv using tab as separator (TabDelimited), there is a chance that the .csv is parsed correctly regardless of the language settings.
Some .csv drivers/providers look into a schema.ini file in the folder containing the .csv file. By specifying the format (Format=Delimited()), you may help the driver to do the right thing.

Related

How to convert HTML file into Framemaker Interchange Format(.mif) file?

I want to mark index and cross-references like Framemaker does.
Framemaker can export the .fm into .htm and .mif file.
I have analyzed how the index and cross-references appears in .htm and .mif file after exporting it from framemaker.
Now my system will produces .htm file and I can manage to mark the index and cross-reference like framemaker does.
I want that framemaker retain the index and cross-references which will be marked by my system.
But there is no way to import or open HTML files directly in Framemaker.
We can import .mif file in framemaker.
So is there any way we can convert HTML files into .mif(FrameMaker Interchange Format).
there is one option, I know its not full proof solution for this problem.
but it can save your efforts to some point,
Save the HTML file to RTF format (using MS word/Open Office)
Open that RTF file in FM
FM accepts the RTF file and convert it into .fm file
Save the .fm file into .mif format
Note : in this conversion, some data loss may happen, i have tried using it for Markers it works but not complete solution.
All the best!!
You can open the .htm files in Structured FrameMaker and then save them to .mif. This will produce less loss in graphics, for sure.

export microsoft export autocorrect acl file to csv

Hello and thanks in advance for any help you can provide. I am a real newbie at all this.
I am trying to export my autocorrected words from Microsoft Word so I can use them in the chrome autocorrect extension 'spelling bee .' The chrome extension allows you to upload files of misspelled words and their autocorrection in csv format.
The problem I face is that my autocorrected list from microsoft word is an acl file. I have spent quite a while trying to figure out how to convert my acl file to a csv file with no success. I thought that I could just manually format the file myself When I open the acl file as a text file, but the formatting and spacing are so off, it would take forever to do it manually.
Is there a straight forward way to open the acl file in simple delimited format? There are many posts online about how to transfer acl files from one system to another, but I could not figure out how to simply convert the acl file into a csv or other appropriately delimited file. If there is a thread out there that addresses this that I may have overlooked please let me know. Thanks again for your consideration.
example of what my acl file looks like when I open it in text editor:
must of had
must have hadmyseflmyselfmyumynaivenaÔvenecassarilynecessarily necassary necessaryneccessarilynecessarily
neccessary necessary
necesarilynecessarilynecesary necessary
negotiaingnegotiatingnkowknownothignnothingnvernevernwenewnwonowobediantobedientocasionoccasion occassionoccasionoccuredoccurred occurence
occurrence
At this link there is a 18kb utitily which will backup your autocorrect to a word document, you can then copy that table into excel and make it into a .csv Autocorrect utility . It's made for Word 97 and 2000, but I used it no problems with Word 2007 and 2010

CodedUI test does not read data from CSV input file

I am having difficulty mapping a CSV file with the Coded UI test method. This is most likely a stupid question but I cannot seem to find a solution for my problem, at least not one that works. I have made sure to set the property of the CSV file to Copy always.
I have also imported the CSV file by writing the following line above the test method.
[DataSource("Microsoft.VisualStudio.TestTools.DataSource.CSV", "|DataDirectory|\\Data\\login.csv", "login#csv", DataAccessMethod.Sequential), DeploymentItem("login.csv"), TestMethod]
The file name is login.csv and it resides in the Data folder.
The test will compile without any problem but once the test executes the fields that should receive input from the CSV file are left empty and the execution is interrupted. I've tried replacing the data from the CSV file by using Strings and it works perfectly fine. The piece of code I am using to import each parameter is:
TestContext.DataRow["Username"].ToString()
Also, the CSV file contains something along the following lines:
Username,Password,Fullname
admin#mail.com,password,Admin
Is there anyone who can point what it is I am forgetting.
Update: I pinpointed the issue, it seems like the issue only revolves around the first column in the csv file. When I try to import any of the other values it works perfectly fine.
Some text files start with a Byte Order Mark (BOM). The CSV reader within Coded UI does not handle the BOM and treats it as part of the first field name. The screen shot below shows the debug trace of a CSV file with a BOM and that same file shown in Notepad++. The DataRow.ItemArray[...] values are as expected. The DataRow.Table.Columns.ResultsView[...] shows the field names, but the first field name includes the BOM.
This CSV file with a BOM was created in Visual Studio using Solution Explorer => Add => New item => C# => General => Text file. Previously I have created a spread sheet with Microsoft Excel and saved it as a CSV file, that file did not have a BOM. I have also created files with Notepad++ and saved as CSV and they did not have a BOM. It appears that Visual Studio creates files with a BOM but when editing CSV files it does not add a BOM.
Visual Studio can create files with the correct encoding. Within "Step 2 - Create a data set" of this Microsoft page it states the text below. (Thanks also to Holistic Developer for providing very similar details in a comment.):
It is important to save the .csv file using the correct encoding. On the FILE menu, choose Advanced Save Options and choose Unicode
(UTF-8 without signature) – Codepage 65001 as the encoding.
For Visaul Studio 2010, i could solve issue be selecting "Western European (Windows) - Codepage 1252" encoding for CSV files.
Summary of steps:
In visual studio 2010, Open CSV file > Go to File menu > Select " Advanced Save Options" > Select "Western European (Windows) - Codepage 1252" > Save.
This should help.
This is not the best solution but its kind of a workaround. I simply set the first element to something random and since I don't need access to the first element it doesn't matter that I don't have access to it.
If anyone finds a correct way to solve this problem I'd be grateful for your solution.

Creating a CSV file with the Report Generation Toolkit in Labview

I want to create .csv files with the Report Generation Toolkit in Labview.
They must actually be .csv files which can be opened with Notepad or something similar.
Creating a .csv is not that hard, it's just a matter of adding the extension to the file name that's going to be created.
If I create a .csv file this way it opens nicely in excel just the way it should, but if I open it in Notepad it shows all kind of characters and it doesn't even come close to the data I wrote to the file.
I create the files with the Labview code below:
Link to image (can't post image yet because I've got to few points)
I know .csv files can be created with the Write to Spreadsheet VI but I would like to use the Report Generation Toolkit because it's pretty easy to add columns and rows to the file and that is something I really need.
you can use the Robust CSV package on the lavag.org forum to read and write 2D arrays to CSV files.
http://lavag.org/files/file/239-robust-csv/
Calling a file "csv" does not make it a CSV file. I never used the toolkit to generate an Excel file, but I'm assuming it creates an XLS or XLSX file, regardless of what extension you give it, which is why you're seeing gibberish (probably XLS, since it's been around for a while and I believe XLSX is XML, not binary).
I'm not sure what your problem is with the write spreadsheet VI. It has an append input, so I assume you can use that to at least add rows directly to a file, although I can't say I ever tried it. I would prefer handling all the data in memory explicitly, where you can easily use the array functions to add rows or columns to the array and then overwrite the entire file.

How to extract hhp file from a chm file

I have an A.chm file for my windows application which runs as expected.
When I decompile it using HTML workshop I get set of html files, .hhc file, .hhk file. When I compile another file B.chm from these extracted files without changing any of the files.((I want to add more html contents to this file but looks like I am losing some information after decompiling)) The output file I get is 72K where as the original file was 75K. B.chm's contents look all file when viewed in the chm viewer but the behavior is lost when when used with the application.
After reading around I found that if .hhp can be extracted from a .chm file then it can be re-constructed as it is without losing any mapping or aliases. Is that true?
How can I extract .hhp file from a .chm file?
Thanks,
Sam
No, Yes , and no.
The original hhp can't be guaranteed extracted
however since chm is an archive type, the project could have added all project files to the archive. I assume you already would have found them if that were the case.
If the decompile process does its administration, it can regenerate the .hhp to a certain degree.
Comments and #define names will probably be lost though, maybe more, but that should not result in problems when recompiling.
But of course it could be that the decompiler is limited. You could try some other (search for something from "keytools").
If not, then take "chmlib" and start drilling down into the format.