When I deploy my app with Jenkins or restart app with app_ctl action_hooks are executed. But when I restart application with rhc app restart app_name actions_hooks are not executed. I set up custom JVM parameters in actions_hooks. What I am doing wrong or how to solve this somehow? Using JBoss EWS 2.0 (Tomcat 7) cartridge.
Try doing an rhc app stop & rhc app start instead of restart, start/stop action hooks are not executed on a restart i believe.
Related
I've trying to write a script to stop Apache and MySQL services in Bitnami WAMP. If we do manually, we have to open C:\WAMP\manager-windows.exe and stop the services like this:
However, when I go into command prompt and run the command "httpd -k start", the apache server starts running and I can access my web app but if I open manager-windows.exe again it doesn't reflect the change. I've been also trying to stop the Apache service using command prompt commands "httpd -k stop" but that is somehow conflicting with Apache web server when it is started through the manager-windows.exe.
Bitnami Engineer here,
You can use the servicerun.bat script we include in the installation directory to start/stop all the services at once.
servicerun.bat START
If you only want to start/stop Apache or MySQL using the command prompt, you can use the scripts we include at installdir/apache2/scripts/servicerun.bat or installdir/mysql/scripts/servicerun.bat
I recently deployed an instance of Ubuntu 16.04 on FIWARE Lab and accessed it using putty, I downloaded docker & docker-compose, I successfully installed fiware-orion & mongo-db as I followed the tutorial, I tried to follow the iot sensor tutorial but whenever I try to start the service it keeps stucking in this infinte loop -> Context Broker HTTP state : 000 (waiting for 200).
Any suggestions?
Details
region:crete
image: ubuntu 16.04
putty infinite loop
The problem was that the docker-compose did not include Orion (and MongoDB) instance which are required dependencies for this tutorial. We have updated the corresponding docker-compose file in order to include both dependencies and now it is working properly. Tips: do not forget to open the corresponding port (3000) in the security and assign a floating IP to the virtual machine to access to the /device/monitor (do not use localhost for accessing it).
I tried to install Context Broker version 1.7.0 on my Centos 6 VM.
I successfully installed it but while the service is running ,when i make a request with curl i get the message "couldn't connect to host"
I checked the port 1026 and the firewall and everything is ok.
Any ideas what may be the problem?
The best way to install Orion is using Docker to test or study: (https://hub.docker.com/r/fiware/orion/).
Docker and docker-compose. You install Orion and Mongodb together, in a quick time. (https://github.com/telefonicaid/fiware-orion/tree/master/docker)
Have you tried using this method?
With it you can use the last stable version (1.10.0).
I'm using google compute engine and have an auto scaling instance group that spins up new VMs as needed all sitting behind a load balancer. I'm also using google's cloud SQL in the same project. The VMs need to connect to the cloud SQL instance.
Since the IPs of the VMs are dynamic I can't just plug in the IPs to the SQL access config so I followed the cloud sql proxy setup along with the notes from this very similar question:
How to connect from a pool of Google Compute Engine instances to Cloud SQL DB in the same project?
I can now log into a single test VM and run:
./cloud_sql_proxy -instances=PROJ_NAME:TIMEZONE:SQL_NAME=tcp:3306
and everything works great and that VM connects to the cloud SQL instance.
The next step is where I'm having issues. How can I setup the VM so it automatically starts up the proxy when it's either built from an instance template or just restarted. The obvious answer seem to be to shove the above in the VM's start-up script but that doesn't seem to be working. So with my single test VM I can SSH into the VM and manually run the cloud_sql_proxy command and all works. If I then include the below in my start-up script and restart the VM it doesn't connect:
#! /bin/bash
./cloud_sql_proxy -instances=PROJ_NAME:TIMEZONE:SQL_NAME=tcp:3306
Any suggestions? I seriously can't believe it's this hard to connect to the SQL cloud from a VM in the same project...
The startup script you have shown doesn’t show the download step of the cloud_sql_proxy.
You need to first download and then launch the proxy. So, your startup script should look like:
sudo wget https://dl.google.com/cloudsql/cloud_sql_proxy.linux.amd64
sudo mv cloud_sql_proxy.linux.amd64 cloud_sql_proxy
sudo chmod +x cloud_sql_proxy
sudo ./cloud_sql_proxy -instances=PROJ_NAME:TIMEZONE:SQL_NAME=tcp:3306 &
I choose crontab to run cloud_sql_proxy automatically when vm start up.
$crontab -e
and add
#reboot cloud_sql_proxy blah blah.
I have created a new OpenShift account for a new application I'm developing.
I have added a MongoDB cartridge for the database, and a Tomcat cartridge for the Java web application.
I now need to connect to the database from my Java web app, but I miss two authentication details:
$OPENSHIFT_MONGODB_DB_HOST
$OPENSHIFT_MONGODB_DB_PORT
As far as I know, I have to type rhc env list -a the_name_of_my_app in the console, but my application seems to have no environment variables set.
What can I do?
Apparently, the default enironment variables are visible only via ssh.
In order to see them, you have to type rhc ssh <appid-as-seen-on-openshift-console> followeb by env.
you can see environment variables by doing ssh to openshift. Also you can use openshift port forwarding feature to setup a connection locally to your database.
Openshift blog link for port forwarding