Version of Newmarket Delphi I'm using:
I am looking for help with creating a CSV from all of the data (Contacts, Bookings) that is available inside of my old Delphi, because I would like to be able to just import the data into our newer Salesforce backed Delphi system without having to input every single booking manually into the Salesforce interface.
Is there a way to do this, or am I just stuck inputting everything by hand?
The input interface on the new system:
Related
I understand that from Business Objects client I have an option to export to "CSV (data only)", but my understanding is that, such an export will not care about the report but just dump the raw universe data.
Isn't there any single way to be able to export the report "view" to CSV ?
It depends on the version of BusinessObjects you're working on.
Originally, the CSV export only looked at the Web Intelligence (I assume you're referring to that particular client) microcube, meaning the raw data retrieved from the data provider(s), and disregards any formatting, filters, aggregations, … you may have specified on your report.
GUI
However, you know have the option to export a report (so not the whole document) as a CSV Archive, which results in a Zip file containing a CSV for the active report at the time of export.
I'm referring to BI 4.1 SP05, previous versions may have this option as I'm not sure when it was introduced.
API
Using the RESTful API that is available in BI4, you can also export a report to CSV. In this case, the actual CSV file will be returned instead of an archive.
Remember that in order to use the RESTful API, you need to have a WACS server in your BusinessObjects environment, running the RESTful API service. You cannot deploy the REST API on an external Java application server.
For more information, have a look at the section Exporting a Report in Listing Mode (SDK information for BI 4.1 SP05).
Remarks
A report is a tab within a document; documents however are often (incorrectly) referred to as reports.
I'd like to import CSV file into Salesforce automatically, e.g. once a day. I can't use Data Loader (Bulk API), because I don't have the needed edition. Is there any other, simple way?
Check to see if DBAmp (on the app exchange) is available for your edition. If it is, then you can use a SQL stored procedure to do it.
Otherwise, I think your only other option would be to schedule an Apex job in Salesforce to grab it from somewhere accessible (like an FTP).
I am trying to export data from a Mind mapping tool such that it can be imported into Rally. Trying to create a mindmap of backlog which can be easily exported to a format compatible to rally ("csv"). I tried using different tools that export data to "csv" format which is Rally-compatible, however ran into some issues and hence decided to get the data into xml format and further convert just the required fields from the xml data to csv which can then be imported to rally.
What I have now is a process that can convert data to xml, I then further use a tool to get it to csv but the hierarchical structure of mindmapping is not maintained when I convert it to csv. So basically, now when I import "csv" data to rally, it doesn't keep parent relationships ( as seen in Mindmapper tool ) in backlogs.
Can anyone help me with this? Thanks!
Have you tried Rally Excel plugin that allows import to Rally from an Excel spreadsheet?
It does not have a direct support for importing parent/child relationship between stories, but if a story designated to be a parent already exists in Rally or imported first, then you may import another batch of stories with Parent field specified, and that will link the newly imported story or stories to the parent. See this video.
I am trying to find the best way to import all of our Lighthouse data (which I exported as JSON) into JIRA, which wants a CSV file.
I have a main folder containing many subdirectories, JSON files and attachments. The total size is around 50MB. JIRA allows importing CSV data so I was thinking of trying to convert the JSON data to CSV, but all convertors I have seen online will only do a file, rather than parsing recursively through an entire folder structure, nicely creating the CSV equivalent which can then be imported into JIRA.
Does anybody have any experience of doing this, or any recommendations?
Thanks, Jon
The JIRA CSV importer assumes a denormalized view of each issue, with all the fields available in one line per issue. I think the quickest way would be to write a small Python script to read the JSON and emit the minimum CSV. That should get you issues and comments. Keep track of which Lighthouse ID corresponds to each new issue key. Then write another script to add things like attachments using the JIRA SOAP API. For JIRA 5.0 the REST API is a better choice.
We just went through a Lighthouse to JIRA migration and ran into this. The best thing to do is in your script, start at the top-level export directory and loop through each ticket.json file. You can then build a master CSV or JSON file to import into JIRA that contains all tickets.
In Ruby (which is what we used), it would look something like this:
Dir.glob("path/to/lighthouse_export/tickets/*/ticket.json") do |ticket|
JSON.parse(File.open(ticket).read).each do |data|
# access ticket data and add it to a CSV
end
end
I have a simple DataGrid that gets the Data from a JSON file hosted on a server.
I am using the build Data Connection Wizard and it works great.
Now I want to let the user manipulate the data and send it back to the JSON file.
I tried to add a form to the data grid but it changes only the data on the application and it is not influencing the file on the server.
any help or direction how should i do that will be appreciated.
Shani
Well i finally used
AMFphp and wrote my own functions.