Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I want to download ALL of google maps(not satellite images just the map) at street level as lets say JPG images labelled as maybe the starting co-ords. need ideas for how to do this? what would be the best language/platform to implement this in? approximately how much space will I need to store it?
Unless Google provides an API to do this, you are on your own to scrape the page somehow and acquire the image. And i don't think its possible to store all the images unless you have a few tera bytes to spare.
if you wouldlike to do this more easily and legally - take a look at openstreetmap.org
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
On my website I want to share a geolocation on which mobile users can click on and then navigation to this location using the users preferred maps app installed on their device.
So I'm aiming to be independent of any maps provider and hence don't want to provide a Link to the API of such. I simply want to encode the longitude and latitude values in a way the user can easily work with.
Any ideas?
You should try using the Geo URI scheme, where a link could look like this:
Map
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I'm making database with movies and I don't know how to save images from Google images and sort them with certain movie.
In database I have 17000 movies and I crawl them from IMDB with import.io but images are too small. Some advice, please?
There is an advanced image search functionality in google images:
You can use that to get the images you want manually, or if you want to automate this with import.io you can append one of the following parameters to a url in a query:
&tbs=isz:l - large images
&tbs=isz:m - medium images
&tbs=isz:lt,islt:xga - images larger than 1024x768
(there are more parameters you can set, this is not a complete list)
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I need to generate a text link cloud something like the image attached.
As some words are vertical, I am thinking of doing it via CSS3. But it is consuming lot of time.
Do you know any website or any better idea of how I can do it fast?
I am using transform property.
A list of websites:
http://www.edudemic.com/9-word-cloud-generators-that-arent-wordle/
http://www.wordle.net/
http://www.tagxedo.com/app.html
http://www.tagcloud-generator.com/
http://tagcrowd.com/
http://www.tagcloudgenerator.com/
Hope this helps! :)
The 0nline Image Map Editor May helps you to create a link
http://www.maschek.hu/imagemap/imgmap
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 2 years ago.
Improve this question
I want to detect the links in page which are used for advertisement. Or is there any statistical data by which I can guess this link is for advertisement?
I know this isn't a concrete answer, but if I were doing the same, I'd take a look at AdBlock and other add-ons in browsers such as Firefox since they do much the same. There are quite a few open source add-ons out there where you can view the code that does this. And even most email programs detect junk mail (and ads) using Bayesian filters which I'm sure with a bit of tweaking would work well.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I am looking at this website - http://www.shopmss.com/ - and I was wondering how they did the splash page, age verification and store all on the same URL 'shopmss.com'. You click through 3 screens before you get back to the store.
My secondary question is, can you do this without setting a cookie? i.e. Javascript, that appends the browser bar URL? Or something with mod_rewrite?
EDIT: I thought this was a relevant question to ask because I was exploring the best practice to accomplish the task, I figured it would have something technical. My bad.
The site is setting a cookie called BX. That could be tracking a session, in which they can display different content based on the state of the session.
They are using a frameset. Check the source.