I am helping remodel a website and was wondering if it was possible to scrape just the Text out of the entire site. Doing a page one at a time using DATA SCRAPER is possible but there is hundreds of pages that need to be worked on. Is there a way to get them all in one scrape? Or further suggestions?
If I understand your question correctly, there is a standalone program called HTTrack (https://www.httrack.com/) that will download an entire website to your local computer. I've used it successfully in the past when I need to grab everything.
Edit: This is my first answer on stackoverflow. Why was this voted down? I'd like to know so I don't do it again.
Related
I am currently editing a webpage I made using Tumblr and it has told me that I have exceeded the 500k character limit. I have nearly one million characters, so I want to take a part of my code, put it onto somewhere else and import it onto my same webpage with a laconic code.
Can anyone recommend me a simple way to do this so that it can all stay on the same page? I have attempted using the iframe feature for larger sections of my website, but it does not perform how I want it to with my CSS. I have seen online that you can use HTML5 imports, but I have seen this feature is highly discouraged and it does not work across all browsers. Thank you for any suggestions!
What I am looking to do is to have a main page with a list of 20 or so photographs on it. When one of the images is clicked, I'd like it to open a new page with the description on that photo, a preview of that photo, and the option to buy it. Is the only way to do this is by writing a new page of code every time a new image is clicked, or is there a way every page can share the same code, while only changing the preview of the photo.
I believe that what you say can't be done with a front-end language, it requires more of a CMS to do such thing, requires back-end job with an engine (PHP as example) and database such as MySQL.
Please remember that the HTML is a Markup-Language not a programming language, it has no logic as "if and else"..
I hope that answered your question. <3
I am developing an universal windows app. I need to download a webpage and extract images from it.
I got the html code and extracted the links to images and downloaded them. The thing is, the site has infinite scrolling (like facebook). When I scroll down to the bottom it loads more images. I am not able to incorporate this into my app. I am a beginner and have very little knowledge of web development or windows app development. This is my first app. I am stuck and have no idea how to proceed. I don't want to use webview as it displays ads from the site and other unnecessary contents. I only want the links to those images. Please help me go past this situation. I need a way to download the new html content that the site loads when user gets to the bottom or some other way to get the image links.
Thanks in advance.
You may or may not be a me to implement this specifically because of the reason you stated. You need to determine how the site loads this information. First I would download Fiddler and in turn enable https connect logging so you can see your encrypted traffic going through Fiddler. Btw the Web View has events you can hook to see loading URLs, etc and it can also be hidden.
So again you need to first understand how the site you want to do this on works and emulate that, assuming they don't have an api already to give you this information as mentioned I'm the comments.
When you do that, come back with code examples and you'll get better help.
I am currently in an internship and making a website for someone. I have nearly finished the website and they would like a way to easily edit the website without messing around too much with code. They have no development team and once my internship is over, obviously they'll be stuck and they would like a way to edit the content on the website. How can I go about doing this? I mentioned a WordPress site to them but they didn't like the sound of that. Are there any other ways in which I can allow them to have this functionality?
Cheers
You can either create a Wordpress-similar, with a login, and while navigating through pages they would have options normal users do not have (e.g: delete, etc).
Or you could create an admin dashboard, where there would be every settings.
I think the second one is the best option for you, as it doesn't require you to modify any pages already created, and you will not have to adapt the style of your website to those supplementary options which would be shown on every pages if you used the first option.
Edit:
Or you could have all the content of the website in files, so that they simply need to edit the files to change the content. Even though I think option 2 is better, this one is a lot simpler to put in place.
I've made a website which displays images hosted on other sites using the html src="http://......" tag, however sometimes some of the images won't load. This appears somewhat random, and I don't think it is a problem with the links themselves.
I display a lot of images, so I am wondering if this is a common problem when trying to load many thumbnails from another site. Is the best solution to host all the thumbnails on my own server, and if so, is there an efficient way to do this (so I don't have to manually download and link to every image)?
Thanks
Is way better to host it on your own server.
Because if are all from other servers, you must connect to all servers and download it.
It causes worse response and increase the time required to load the page.
To the image and links downloading - I think it is possible, just go on google and try to find some advanced html page downloader. I had one and it worked directly the way you want. - can't remember the name..
(also sorry for my bad English)