Retrive and process Ckan Data using Fiware Platform - fiware

I'm finishing the publishing of data in Ckan and I want to retrieve, process and visualize this data using Fiware platform. But I searched for this and couldn't find any document showing this integration. Could anyone help me with processing and visualization of Ckan data using Fiware platform?

You can use SpagoBI for data visualization and analysis.
Here you can read how to integrate SpagoBI with CKAN.

Related

Watson training data format

I want Watson to train on certain data the user will provide in my web app - the data will be posted through forms.
My question is - which service in IBM Cloud matches this the best? I've tried Discovery, but it doesn't seem to be the best match for my request, first of all it does not want to accept .json or Excel formatted files, which seems like a red flag to me (concerning what I am looking for).
My ultimate goal is for Watson to learn the patterns and ultimately start providing suggestions for the user.
My data I give to Watson would look like this, .json format:
{
"songName" : "Beyond the sea",
"artist" : "Bobby Darin",
"genre" : "jazz"
}
Thank you in advance.
I've setup my IBM Cloud, enabled Discovery as a service, attempted to upload .json and excel files, from which both have been rejected.
I expected Watson to process the provided structured data, find patterns, and provide intelligent suggestions.
If you are planning to build and train a model, then Watson Machine Learning service on IBM Cloud is what you are looking for. After you build and train a model, you deploy it and put it into production, so you can pass data to the model and get scoring data, also known as predictions, back.
You can access your deployment through an API endpoint, using the Watson Machine Learning Python client, CLI, or REST API (manually or in an application) to analyze data.
To know more about deploying your model, refer this the documentation here
Also, there is a code pattern that serves as an example of how to build a Machine Learning recommendation engine using Jupyter notebooks on Watson Studio

Retrieving JSON data from RESTful API to Tableau

I have a RESTful API that I hosted on Heroku:
https://newsaggrarab.herokuapp.com/getSpecCountries
This retrieves me my data from my MongoDB mLab database in JSON format and I was able to use it just fine in Qlik Sense to create visualizations. However, I can't connect to this API in Tableau Desktop and haven't found any helpful guides on the internet. I just want to retrieve this data and use it in my Tableau Desktop visualizations.
Thanks in advance
Tableau implemented a 'Web Data Connector' in version 9 which might be a good route for you. It requires that you write some code (JavaScript, I believe) in order to create your own Web Data Connector to bring in your JSON data.
Here is a great resource to start your search. Here is a link to the WDC community. There are plenty of examples and even a code simulator.

OpsCenter backend API for generating custom reports

I would like to know if we have any specific api's in OpsCenter which would help in generating custom reports.
The use case is:
I am consolidating reports from different tools like App Dynamics,OpsCenter and others into one csv or excel sheet for which i would require any opscenter api which would allow me to do a backend call which would pull the required info for me from OpsCenter and then i can do some additional stuff and consolidate it along with other data in a csv or excel in an automated way.
Any api which would help me with backend Opscenter calls for reports data would be helpful.
The OpsCenter API is fully documented at http://docs.datastax.com/en/latest-opscenter/opsc/opscApi_g.html
If the kinds of reports you're looking to generate involve metric data, you'll be particularly interested in the new-metrics endpoint: http://docs.datastax.com/en/opscenter/6.0/api/docs/metrics.html#method-get-new-metrics
Otherwise you'll just have to wade through it to find what you're looking for. But the front-end is all driven through those api's, so anything you see in the web-interface can be replicated via the documented apis.

Any ESB frameworks which provides API to host a webservice dynamically

i am performing an R&D to check whether are there any open source ESB frameworks which provides the following features
1. API to host a SOAP webservice/consume a service by providing the required data. We are working on a tool to design webservices for our product and the metadata created from the tool would be input to the API. What we are looking is that the metadata generated should be hosted as a webservice instantly without restarting the ESB container.
2. API provided by the framework to define the count of a service executions performed.
3. ESB framework which supports versioning both in hosting a service and also consuming a service.
It would be very helpful if someone can provide a direction to such ESB frameworks
Talend ESB provides the functionality you described.
it can consume a WSDL (your metadata created from your tool) to quickly create a web service service. It is hot deployed into a Karaf container (i.e. do not need to restart the ESB container)
you can use built in logging or custom logging to count executions performed.
it can be hooked into git or svn for versioning.
There is a free version you can download here.
Here is a basic tutorial on the product which demonstrates a. ingesting a WSDL, b. hot deployment, and c. execution count (though it does not go into how to log that).

How to expose mysql data as a soap web service in Talend

I want to expose data in a mysql data source as a soap web service using Talend. How can I achieve that. Please advice.
The most powerful way to do so to do so is probably to build a WSDL using Talend Open ESB Studio (Tutorial)
Alternatively, if you don't like to use another piece of software or you prefer a quick-and-dirty solution, you can achieve this with standard Talend Open Studio for Data Integration. You just need to draw a standard Talend job which redirects your mysql data output to a tBufferOutput instance. Then, you must export your job as Axis webservice, finally deploy it in your application server (Tutorial). This way, your buffer data will be automatically exposed via web service. But it has a drawback: you cannot tune the response (ie, add labels to fields or refactor the SOAP response structure at all).