I want to deploy DAML to QLDB. I've read the documentation and issues, but it doesn't write how to deploy to QLDB.
I read this blog, Can't I still use DAML on QLDB?
https://blog.daml.com/daml-driven/quantum-daml-amazon-qldb-goes-ga
Blockchain Technology Partners have documentation how to deploy DAML to QLDB here: https://blockchaintp.com/sextant/daml/getting-started/#daml-on-qldb
Related
I am new to devops and reading openshift docs about this. Seems both buildconfig and pipeline(tekton in openshift 4.6) can achieve source-to-image process and triggered by git webhooks. So what is the difference between openshift buildconfig and pipeline?
PS:
Just finished the pipeline tutorial on openshift, there is no build or buildconfig resource created during the whole process.
Openshift buildconfig is "Openshift specific" and was very hot in Openshift3.
The hot stuff then was the source2image thing.
Buildconfig could be setup for S2I, Docker and even "Pipeline". But this is not to mix with Openshift Pipelines with Tekton. The BuildConfig pipeliens was provided using jenkins files.
Now as Tekton has gain more stability, respect and maturity out in the community and also under the "Openshift Pipeline" it has been the right way to do stuff.
It is a more complete way to setup complex pipeliens with k8s native way and not only for openshift.
So what the difference more than above I would say that using a pipeline will give you all the flexibility and power as any CI build tool. It is frequently updated and has a great slack community. Buildconfig has lot of limitations on what you can do.
All you can do in buildconfig and more is achievable in Tekton pipeliens, but not the other way around. ;)
When using Openshift Pipelines there are tasks provided for s2i too:
https://github.com/openshift/pipelines-catalog
Also Tekton tasks can be added from:
https://github.com/tektoncd/catalog
I tried to upgrade DAML SDK for purpose of testing, but now I want to downgrade version for running code already written for 0.13.36 version.
DAML SDK version is controlled project by project by setting the sdk-version in daml.yaml. So if you have a project written with SDK 0.13.36, just make sure the daml.yaml still contains sdk-version: 0.13.36. daml start and daml build will then do their job using that SDK version.
Is there any link to find OpenShift 4.2 Rest API documentation?
I could see latest that we have is for 3.11
https://docs.openshift.com/container-platform/3.11/rest_api/index.html.
Will is absolutely right, and here is the link to doco:
https://docs.openshift.com/container-platform/4.3/rest_api/index.html
When in doubt, this command is your best friend
oc --loglevel=9
OpenShift Rest API documentation is expected to be available once 4.3 is released. In the meantime, OpenShift 4.2 is built on top of Kubernetes 1.14; the 1.14 Rest API documentation is available at https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.14/.
For OpenShift specific resources, if you're looking for the resource definitions, you can use oc explain <RESOURCE> --recursive from the cli to see a full resource definition.
I'm trying to replicate the installation guide to integrate Wirecloud 1.0.1 with the latest version of IDM Keyrock.
I'd like to understand if the guide in the documentation is valid again for this version of WC.
Thanks in advance for the answers.
As you suspected, the linked documentation is not valid for WireCloud v1.0.1, you linked to the latest branch of the documentation, that is the documentation for the development version of WireCloud.
The documentation on how to configure WireCloud v1.0.1 to use the IdM can be found in this link.
We know that is not clear that the latest branch is the one used for document the development version, but Read The Docs does not allow us to rename (or we do not know yet how to do it) this branch using a more meaningful name (e.g. develop). The default branch is stable (pointing to the documentation of the latest stable version of WireCloud) but it has proben to not be enough. We have created a ticket for improving this aspect.
The Kaa platform as an IoT cloud platform is prebuilt to run on amazon aws or a virtualbox sandbox. Is it immediately deployable to openshift, especially the starter free plan? If not, what it takes to get it to work?
I have looked at the python on openshift which uses the S2I to dockerize a software collections version of python, e.g. 2.7. I'm wondering how these projects or technologies would work together to make Kaa to run on multiple platforms, or to make more versions/flavors/variants of Kaa to run on platforms. An interesting question, but I'm not sure the way of thinking is right. Though this is just to add some hints of the background information that I've been looking at, may or may not be related to the questions asked here.
You can use different workflows to achieve the goal:
It seems there are Docker images ready to use[1], so you can try deploy it in Openshift and see what happens.
You can create a custom s2i[2] image in Openshift and create a Dockerfile with all the base software you need to run Kaa.
You can create a Dockerfile (maybe editing the exsisting Kaa Dockerfile) that contains/add all the software you need, then create a BuildConfig with docker strategy[3] and run it in an Openshift project to add your Kaa image to the imagestream and then deploy Pods from your Kaa imagestream with a Deploymentconfig[4].
[1]: https://kaaproject.github.io/kaa/docs/v0.10.0/Administration-guide/System-installation/Docker-deployment/
[2]: https://blog.openshift.com/create-s2i-builder-image/
[3]: https://docs.openshift.com/container-platform/3.7/dev_guide/builds/build_strategies.html#docker-strategy-options
[4]: https://docs.openshift.com/container-platform/3.7/dev_guide/deployments/how_deployments_work.html#creating-a-deployment-configuration