Im currently trying to import json data into elastic search.I have installed elastic search v2.2.1 in windows7 & trying to import json data into elasticsearch. But Im unable to import the same. I have checked various posts and they are specifying to execute the curl command. Below are the steps that I have tried
Tried to execute the below curl command via command prompt.But got a message that 'curl command is not recognised as an internal or external command'.
Installed cygwin and tried to execute the curl command, got a message'-bash: curl: command not found'.
I also tried to install sense chrome plugin, but unable to download the plugin and found that the plugin is no longer available.
I also installed the inquisitor plugin for elastic search, but looks like it helps to validate the query and does not help in importing the data into elastic search.
Any help on this is much appreciated.Thanks in advance.
For curl in windows you could try the download link here
The other way to use in windows is to use the Sense, which is a Kibana app. You can install it from here
The Sense app replaces the Chrome plugin
Related
I want to import and use dataset package of python at AWS Lambda. The dataset package is about MySQL connection and executing queries. But, when I try to import it, there is an error.
"libmysqlclient.so.18: cannot open shared object file: No such file or directory"
I think that the problem is because MySQL client package is necessary. But, there is no MySQL package in the machine of AWS Lambda.
How to add the third party program and how to link that?
You should install your packages in your lambda folder :
$ pip install YOUR_MODULE -t YOUR_LAMBDA_FOLDER
And then, compress your whole directory in a zip to upload in you lambda.
What you have to do is to include the binaries needed with your lambda package.
You need to utilize pip and create an isolated environment.Your zip uploaded to lambda needs to have the python2.7/site-packages included (the ones installed with pip).
Now there are extreme cases of os-related dependencies.
This has a tricky solution.
In those cases you have to spawn an amazon linux ec2 instance in order to build/get those dependencies and package them with your lambda.
Once your lambda is packaged you can close the ec2 instance.
Check this guide if virtualenv is not enough.
This is an os dependent system file. I'm guessing that you successfully installed the Python mysql client, but you still need the system mysql client, which seems to be a different version on your system than the lambda one. While building your virtual environment on the official lambda image will definitely fix this problem, you might have some luck copying your own copy of this system file into your lambda zip file.
I found mine with
locate libmysqlclient.so.18
Note: depending on your system, the version number at the end might be different. Use the version in the error you receive.
Adding that file on the top level of my zip file with
cd \path\from\locate\to\libmysqlclient
followed by
zip -u \path\to\lambda\zip\file.zip libmysqlclient.so.18
worked for me.
I installed a cakephp 3 app via composer from the root of my website c:wamp/www/my_app, and I developed untill i need to implement acl.
I installed the acl plugin for cakephp 3 via composer for the same path c:/wamp/www/myapp and i configured it, was everything well, but when i test the implementation of acl, i get the following error:
Missing Component, Create the class AclComponent below in file: src\Controller\Component\AclComponent.php
and seems to be true, theres no such file in the above location, instead the file is in vendor/cakephp/acl/src/controller/component.
Was investigating and i find that composer doesnt install plugins in a cake default path, theres many articles about, but how doi correct my installation? I already did all the stuff to implement the access control lists.
Thanks in advance error screenshot
I had an py27 application running on aws ElasticBeanStalk for over a year. Recently it stopped working. So I tried to redeploy. During redeploy I got the following error in the logs:
Return code: 1 Output: [CMD-AppDeploy/AppDeployStage0/AppDeployPreHook/03deploy.py]
command failed with error code 1:
/opt/elasticbeanstalk/hooks/appdeploy/pre/03deploy.py
New python executable in /opt/python/run/venv/bin/python27
Not overwriting existing python script /opt/python/run/venv/bin/python
(you must use /opt/python/run/venv/bin/python27)
Can somebody tell me where I will find /opt/python/run/venv/bin/python? How can I change it to
/opt/python/run/venv/bin/python27?
Thanks
When you redeployed, you opted into the latest beanstalk version, which uses a different AMI than the one originally used. If you're familiar with the concept of ebextensions then you're probably looking to add a file like this:
.ebextensions/python.config:
ln -s /opt/python/run/venv/bin/python27 /opt/python/run/venv/bin/python
I have a working vagrant + ansible setup to provision my digital ocean, it was running on api v1, but when DO deprecated it I got an error message telling me there was no support for v1 anymore. After a research I found out I needed to upgrade to ansible 2.0 + update my digital_ocean.py since the older one was still using client_id and api_key, the new one however now uses the api_token.
Basically I've updated
1.digital_ocean.py which I got from ansible repo module
2. digital_ocean.ini to contain the api_token
3. updated my api token from DO to make sure it's using the new one
but when I execute my ansible playbook I initially got this error
ERROR! The file provisioning/inventory/staging/digital_ocean.py looks like it should be an executable inventory script, but is not marked executable. Perhaps you want to correct this with `chmod +x provisioning/inventory/staging/digital_ocean.py`?
So naturally had to chmod +x it but when I did I get a new error which is
ERROR! The file provisioning/inventory/staging/digital_ocean.py is marked as executable, but failed to execute correctly. If this is not supposed to be an executable script, correct this with `chmod -x provisioning/inventory/staging/digital_ocean.py`.
ERROR! Inventory script (provisioning/inventory/staging/digital_ocean.py) had an execution error:
ERROR! provisioning/inventory/staging/digital_ocean.py:3: Error parsing host definition ''''': No closing quotation
The next one seem to be json parsing related, my only problem is that it's on line 3 which if you check on the code itself are still on the comment side
https://raw.githubusercontent.com/ansible/ansible/devel/contrib/inventory/digital_ocean.py
I received both errors mentioned and solved them with the following:
The dopy related error was due to the fact that my python and pip were installed differently. dopy was installed via pip which was installed via homebrew. I was using the system python. When I installed python via homebrew, the script found dopy just fine.
Regarding the second error, that is the result I got when not setting the DO_API_TOKEN. I set mine in the command itself with:
DO_API_TOKEN=<api_token> ansible -i digital_ocean.py all -m ping
I am creating a registration form, i want to add a google captcha to the form however i noticed that the package requires PHP 5.4 or greater and "ext-curl" and "ext-json" i have looked around to try and find out how to get these dependencies installed but i have found no solution, currently when i try adding the google captcha package to my composer.json and doing a composer.update i get an error then i have to take it out from my composer.jon file and re-run composer update to fix the error. I think it might have something to do with these 2 dependencies does anyone know how i can get them installed?
I use the inbuilt PHP server in Laravel on windows 7 platform.
Directly from easyphp administration, use test code:
<?php phpinfo();?>