how to set-up hadoop cluster multi-node on ubuntu? - hadoop2

I want to install a multi node hadoop cluster. I want to have name node on one machine, job tracker on another, data nodes on seperate machines. Do I need to install hadoop on each of these machines and run some specific agents to run the name mode on machine a, hdfs on machine b, job tracker on machine c and so on. How do I go about setting this up?

Related

Windows Service installation on Docker Container

I am trying to port one of our Tomcat based Windows application to Windows Nano Server Container. The installation process includes two Windows Service installations - an ActiveMQ service and another Tomcat Service. This is achieved using two batch files provided with the product. In addition to installing the service, they also pass several arguments pertaining to JVM. Now the issue I am facing is I am not able to get these services installed on a Nano Server. The same Dockerfile works perfect and installs the product on Windows Server Core Container.
Observation: The services which get installed through batch files are dependent on two system components - Ancillary Function Driver for Winsock (Service name: afd) and TCP/IP Protocol Driver (Service Name: tcpip). This is something which I discovered on a VM where the product was installed when checking the service properties. I see that they are absent in a Nano Server Container but are present in Windows Server Core container. Could that be a reason? If yes, is there any way to get those services running on Nano Server?
I am at a loss on how to troubleshoot this issue. Are there any limitations on Nano Server Container when it comes to Windows Services?
Consider building your docker container from the ground up,
Determine and set your prerequisites
Install all of the dependencies first
Install the application and services in question
(For additional information you can look at this wiki page.)
The Nano server container is minimal and misses many windows default services, you’ll have to build most of them through Dockerfile yourself.
The easiest method will be to do this manually on Nano server and confirm everything to be working in the docker container/image, and then later generate an automated build.

Should I install MariaDB to my git repo, or locally / individually to all development computers?

I am starting work on a website that will use MariaDB for storing information (no sensitive information), and would like to keep everything in my git repo.
Originally, I planned on installing MariaDB to the separate computers that I plan to develop on (my desktop and laptop), but decided that it may be easier to store all of MariaDB (the program and the databases) in the git repo, so that one would just need to clone the repo and run MariaDB from the repo just like they would run Node, but I have not found any information on how to do this.
My questions are (1) should I install MariaDB and its databases to my git repo, instead of installing MariaDB in /usr, and the database in /var/lib/mysql, and (2) how would one do that?
Instead of attempting to put a mariadb runtime environment inside your version control, consider using docker to describe how to configure an appropriate mariaDB installation. I use makefiles atop that to contain the commands I use to build and run the docker but you coould just as easily use a shell script. Finally, provide a dataabase load script that loads your test database from a text file within the repo.
using docker to describe runtime environments for your application and dependencies is awesome. It strikes a great balance between having an incomplete git repo, and having to put binaries and database data in your version control. You wouldn't want to track changes to the underlying maria db files, anyway, so best not to commit them. You can build the docker containers you need on every workstation you use without much trouble, your automation around creating them provides a mechanism to ensure consistency, and by loading a database with the correct test data every time you develop your app, you'll have a better development process and less schema and data related changes. It works great, I do nothing but docker driven development these days.

Switching between developer environment Windows/Ubuntu

I have a PC with a dual boot: Win 8.1 and Ubuntu 14.04 and I have some project in PHP & mySql and sometimes I need to develop under ubuntu (but I usually prefer Windows) but I need to share the files and db's between the two os.
In share I mean if I change some php files or rows in the database that must be visible in both OS. Is it possible?
Extra thing: The most beautiful thing would be if I could get the mercurial in this thing.
Liquibase for proper versioning of DB-changes in both OS
Mercurial in both OS
External Mercurial hosting (Bitbucket?) for intermediate repository and backup
Good editor or IDE, in which you can configure EOL-style (in order to have same EOL in Windows and Linux, different by default)
Code in active OS, commit changes to local Mercurial, push to remote.
Before starting work in another OS - pull collected changes from remote

What do I have to be aware of when installing Reporting Services on a productive SQL Server 2008

I have a running sql server installation on which I would like to install the reporting services module. Now because it is very important that the sql server it self keeps available and running without any interruptions I get a little scary about this step.
Now my question. Are there any pittfalls I have to be aware of before I install the module? E.g. does the server restart during the installation?
Can you try this change in a virtual machine first?
For example, use the Disk2vhd application (http://technet.microsoft.com/en-us/sysinternals/ee656415.aspx) to make a snapshot of your current production system. Then, use the VHD that was created in Microsoft Virtual PC, Hyper-V, or Windows 7 native VHD boot.
If you try a virtual machine for testing, disable networking or create a private network before starting the virtual machine! Having two servers with the same name and SID on your network is a Bad Thing.
(If you're not using virtual machines, isn't it time? :)

Remote installation of Stored Procs on MySQL

I'm just setting up the live environment for my new project. Unlike my dev and testing systems, the live environment consists of a web server (Win 2003) and a separate DB server (MYSQL).
My installation process for each release of the software is nicely scripted, giving me full rollback options etc.
However, I can't work out how to install my stored procedures within that process. I can't run a MYSQL command line because MYSQL isn't installed on the web server, it only accesses the DB via ODBC.
Is there a means by which I can run MYSQL commands on the web server, via ODBC from a command line? I really want to keep it all together so I can run "Install v123" and everything whizzes off and gets installed in one go.
There may be a more elegant solution but: I had a very similar problem a number of years ago, and I eventually just wrote a little stand alone program to run my scripts at the end of the install.
Another common option is to have them run as part of a configuration utility/page the user goes to after setup, but I'm assuming you want to keep this as a 1 step installation.