I am trying to log in to the Bluemix Container Service with cmd.exe. I installed the latest Bluemix CLI and IBM Bluemix Container Service plug-in, and tried to login using the following command:
C:\>bx login -a api.ng.bluemix.net
...
OK
...
C:\>bx cs init --host https://us-south.containers.bluemix.net
The IBM Bluemix Container Service Kubernetes beta in the US South region (Dallas) has concluded. You can access your bet
a clusters via the CLI using (bx cs init --host https://us-south-beta.containers.bluemix.net). See https://ibm.biz/cont
ainer-migration for beta migration information.
Using user provided API endpoint: https://us-south-beta.containers.bluemix.net
OK
C:\>bx cs clusters
The IBM Bluemix Container Service Kubernetes beta in the US South region (Dallas) has concluded. You can access your bet
a clusters via the CLI using (bx cs init --host https://us-south-beta.containers.bluemix.net). See https://ibm.biz/cont
ainer-migration for beta migration information.
FAILED
The client is not yet configured. Run `bx login`.
The Bluemix login was successful, but an error occurred. I tried it with bash on Windows, but the results were the same.
I would recommend to install Cloud Foundry CLI instead of Bluemix CLI and work hope your problem gets resolved.
Related
I have a openshift cluster on IBM cloud. I want to connect to the worker nodes using SSH via Putty but documentation says,
SSH by password is unavailable on the worker nodes.
Is there a way to connect to those?
If you use OpenShift v4 on IBM cloud, you may access your worker nodes using oc debug node/<target node name> instead of SSH. oc debug node command launches a temporary pod for the terminal session on the target node. You can check and run linux commands like usual SSH session through the Pod. Try it.
SSH access to worker nodes in OpenShift is disabled for security reasons. The documentation suggests to use DaemonSets for actions to be performed on worker nodes.
I am user of Openshift online and OKD. I am facing similar issue in both places. Please have a look.
I have created a project.
I have launched php in Developer's Catalog option. With other details, I entered my project's git url, project is cloned successfully. Now it needs to connect to mysql database only.
In Pods, I deployed mysql image from 'Deploy Image' option. It is launched successfully.
When I make mysql connection from php pod to mysql pod, it does not connect, connection time out.
How should I make connection?
Note :
I do not have datastore option to launch mysql from developer's catalog in openshift online, that's why I am launching mysql image from deploy image.
As you mentioned you are using Openshift Online and OKD and you are facing the issue at both places.
You can not create mysql from development store because currently, the OpenShift Online catalog does not provide MySQL template via the web interface directly, but you can deploy the MySQL template using the oc CLI instead. The database deployment is simplified when using templates.
Once logged in with the oc CLI, running
oc new-app -L
will list all of the templates that we were used to seeing in the web console, including the mysql-persistent. Then, you can specify all the template parameters via the oc CLI, e.g.:
oc new-app mysql-persistent -p MYSQL_USER=<desired_DB_username> -p MYSQL_PASSWORD=<mysql_password> -p MYSQL_DATABASE=<desired_database_name>
If you'd like to see all the supported template parameters, you can use
oc process <template_name> --parameters -n openshift
or, for a more detailed output,
oc describe template <template_name> -n openshift
Once the app is launched successfully, you can find this app's hostname in services and connect to it from your php pod after defining host name in php configuration file.
I recently deployed an instance of Ubuntu 16.04 on FIWARE Lab and accessed it using putty, I downloaded docker & docker-compose, I successfully installed fiware-orion & mongo-db as I followed the tutorial, I tried to follow the iot sensor tutorial but whenever I try to start the service it keeps stucking in this infinte loop -> Context Broker HTTP state : 000 (waiting for 200).
Any suggestions?
Details
region:crete
image: ubuntu 16.04
putty infinite loop
The problem was that the docker-compose did not include Orion (and MongoDB) instance which are required dependencies for this tutorial. We have updated the corresponding docker-compose file in order to include both dependencies and now it is working properly. Tips: do not forget to open the corresponding port (3000) in the security and assign a floating IP to the virtual machine to access to the /device/monitor (do not use localhost for accessing it).
I am a newbie to Bluemix and have problems with containers. I installed Docker Toolbox (my OS is Windows 7). Afterwards I installed Bluemix plugin IBM-containers. (I checked both installations, everything seemed fine to me.) Then I logged into Bluemix (cf login -a ...). And then I run command: cf ic login. And I got an error message
I've tried to deploy a sample java application to Bluemix before. It worked OK, I did not encounter any network connection problems. I don't understand what is the cause of the error.
Any ideas what could be the problem or how to fix it? Thanks in advance.
You need to login using:
ice login -k API_KEY -R registry-ice.ng.bluemix.net
Source: Alaa Youssef on DeveloperWorks
If this is the first time you use the container service you may need to run the cf ic init and the cf ic namespace set commands. Take a look here for more information.
If your registry is already configured, please always remember to do a cf login just before doing a cf ic login. If the problem persists and you are able to access containers service from your Bluemix Dashboard, you could try to use the new bx CLI (that wraps the ic plugin).
I have created a new OpenShift account for a new application I'm developing.
I have added a MongoDB cartridge for the database, and a Tomcat cartridge for the Java web application.
I now need to connect to the database from my Java web app, but I miss two authentication details:
$OPENSHIFT_MONGODB_DB_HOST
$OPENSHIFT_MONGODB_DB_PORT
As far as I know, I have to type rhc env list -a the_name_of_my_app in the console, but my application seems to have no environment variables set.
What can I do?
Apparently, the default enironment variables are visible only via ssh.
In order to see them, you have to type rhc ssh <appid-as-seen-on-openshift-console> followeb by env.
you can see environment variables by doing ssh to openshift. Also you can use openshift port forwarding feature to setup a connection locally to your database.
Openshift blog link for port forwarding