Add Unity 300 to zabbix - zabbix

I would like to ask about adding Unity 300 to Zabbix. I have done some research on this issue and I have found out that Unity doesn't support installing zabbix agent and snmp. Only snmp trap is allowed. I have added Unity to zabbix and I am able to ping it. I have activated snmp trap and it is working. I was also able to get some metrics with uemcli commands with ssh connection. But is not enough. I have found some templates for Unity 600 but it needs zabbix agent. aklyuk/zabbix-emc-unity. I need more metrics and in another format. Is there another way to get metrics?

You need to use Unity REST API to retrieve metrics. Look for the specs on the EMC KB.

Related

Legacy GCE and GKE metadata requests from google_daemon/manage_addresses.py

I have an old Debian Compute Engine instance (created and running since December 2013) and got an email warning about the turndown of Legacy GCE and GKE metadata server endpoints (more details at https://cloud.google.com/compute/docs/migrating-to-v1-metadata-server).
I followed the directions for locating the process and found that the requests were coming from /usr/share/google/google_daemon/manage_addresses.py. The script seems to be the same as what's at https://github.com/gtt116/gce/blob/master/google_daemon/manage_addresses.py (also with what's in that directory).
I don't recall installing this, so I'm imaging it came with the provided Debian image I used in 2013.
Does anyone know what this manage_addresses.py script is, what it does, and what I should do with it now that the legacy metadata server endpoints are turning down? Is it safe to just stop running it? Or is there a new script I should replace it with? Or should I just try to update it myself to use the new endpoint?
I dug around and was able to trace /usr/share/google/google_daemon/manage_addresses.py as being installed by a package called google-compute-daemon. A search for that brought me to https://github.com/GoogleCloudPlatform/compute-image-packages#troubleshooting which explains that google-compute-daemon has been replaced with python-google-compute-engine. That led me to https://cloud.google.com/compute/docs/images/install-guest-environment . I followed the instructions there and manually installed the guest environment.
I noticed during installation that it said it was removing the google-compute-daemon package (and a packaged called google-startup-scripts), so this seems like the right thing. And I'm no longer seeing any requests to the legacy endpoints. So it seems like at some point the old guest environment failed to update.
TLDR; If you have this problem, follow the instructions at https://cloud.google.com/compute/docs/images/install-guest-environment#installing_guest_environment to manually update the guest environment.

How we can monitor a service status using Zabbix?

We are using Zabbix for server monitoring and its working fine for system resources like disk, CPU, memory etc.
Now we want to monitor some services also whether they are running fine or not like Apache, Nginx, Puma, Sidekiq etc.
Can you please help me how we can monitor such services using Zabbix?
Any guidance will be appreciated.
Thanks in advance.
You should refer to the documentation, it covers windows service monitoring and generic process monitoring with proc.* items.
Here you can find the supported item by platform matrix.
There's an external template for systemd lld, you can find it on Zabbix Share
for Nginx monitoring you can use that template
also take a look this repository, probably you can find there something useful
For sidekiq specifically, using
proc.num[,,,sidekiq]
seems to work. It uses the cmdline -argument.
Source:
https://zabbix-users.narkive.com/EKVrN9VY/proc-num-item-for-sidekiq-process

Setting up chrome remote desktop on google cloud compute vm linux instance

I'm trying to set up GUI access to a linux VM on google compute engine. I've followed the advice here:
https://groups.google.com/forum/#!topic/gce-discussion/tN9oZs8xWps
I can get as far as getting the instance to show up on the "My Computers" section of the CRD app, but it is grayed out.
I'm wondering if I need to mess around with firewall settings, or try a different desktop (I've been trying cinnamon). I'd appreciate any help. Thanks!
You definitely have to open the ports to connect, so you will have to use a firewall rule. Cloud Platform has an implicit deny ingress rule. It is explained in the thread you posted that also if you are using CentOS there is an additional steps to disable a firewall rule.

OpenShift system and package updates/patches

How does one keep OpenShift gears up-to-date? For example, updates to:
The Linux kernel
Important components/libraries like libc
Apache
Apache modules like mod_wsgi
Python
Python packages
Does OpenShift automatically update these and then restart the gear (or reboot the node)? Or does OpenShift send email notifications and the end-user can restart the gear during maintenance windows? What is the model?
What got me thinking about this was back in January there was a remote-code-execution bug in Ruby on Rails that everyone had to patch immediately.
This FAQ seems to suggest that some level of upgrades happen automatically, but it isn’t clear whether this only applies to the OpenShift-specific code, or also other components like the kernel, Apache, etc.
I can tell you from my experience that changes to the openshift system are not always automatic. They had a change about 10 days ago and I'm still tracking down what they did to make my app run correctly. As far as I know there was no email sent. I did find a blog post of some of the major changes, not all. Of course, they introduced at least one bug that I know of. YMMV
My experiences over the last few weeks have been the following:
Last week there seemed to be an unannounced reboot of the server. I detected this by logging from a custom action hook. I didn't receive any email about it and I didn't see any notice at https://twitter.com/openshift_ops or https://openshift.redhat.com/app/status.
This week, there was the Heartbleed OpenSSL vulnerability and it seems like some gears were restarted. I didn't receive any email about it, Twitter didn't show anything, but there was information on the status page.

Server Monitoring tools Apache/MySQL

my boss has asked me to find a tool that will monitor our sever health. Some kind of desktop application preferably that we can keep an eye on and will monitor us when capacity goes over a certain level, or we approach max storage etc.
We need to monitor both MySQL and Apache. I'm guessing I might need two tools.
THanks in advance
Have you looked at munin? it's not desktop... but i don't know why do you want to have a desktop solution?
you can monitor apache with SNMP module like mod_apache_snmp and tools like OpenNMS, and Nagios. Nagios supports monitoring mysql also.
You might also like to look at Megamon (http://www.megamon.com). Megamon is a complete monitoring solution capable of graphing numerous system performance aspects as well as escalating alerts and much more.
Megamon is not a desktop solution but runs as a Virtual Appliance. However, since it has a web interface, it is just as easy to use as a desktop application.
Have a look at SeaLion. SeaLion is a cloud based Linux server monitoring tool. Getting started is as easy as executing a command. It installs an agent at /usr/local/sealion-agent and runs as an unprivileged user (sealion). This agent will collect data at regular intervals across servers and this data will be available on your workspace. The latest version is shipped with NGINX, Apache, MySQL, MongoDB and Redis monitoring capability. It is free for 1 server with a 12 hours data retention policy.