This question already has answers here:
Storing Images in DB - Yea or Nay?
(56 answers)
Closed 8 years ago.
I'm making a website for my brother's webcomic, which was previously hosted on Tumblr. What is the most efficient/logical option for storing the pictures?
Downloading and putting the path in the Db
Storing them in the database, base64-encoded
linking directly to the pictures on Tumblr
wat do?
If the tumblr site is going to remain active I would lean towards using the Tumblr API to get at the photos. You could then just write some javascript/jquery functions to display the images however you want.
I've done something similar in the past with Google Picasa Albums and it worked out pretty well.
http://www.tumblr.com/docs/en/api/v2#photo-posts
Just a little additional info, in the past I've found using jquery plugins sometimes makes it a bit mor simple to get at the data I'm looking for.
Never used this one in particular but just a quick search and found this as an example of one that might be helpful.
https://github.com/Iaaan/jQuery-plugin-for-Tumblr-API
Related
This question already has answers here:
Is there anyway to render a website and share it without hosting it?
(2 answers)
Closed 1 year ago.
So recently I started a project to expand my knowledge when it comes to HTML and CSS. I made a little website just for myself and I want to get some feedback from my friend.
I want him to be able to maybe click on a link and then get to my website (doesn't have to be a link). So I don't want to publish the website for anyone but for my friend or anyone that has that link or something like that.
I already heard about sharing files to someone on OneDrive and including all the code in it but my OneDrive is already out of storage so something different would be nice.
It is best to use pages.github.com check the page is a small post and video tutorial on how to use it, greetings bro!
This question already has answers here:
Pros and Cons of a separate image server (e.g. images.mydomain.com)?
(5 answers)
Closed 6 years ago.
an example of what i am asking, the website
www.discogs.com
has all of its image paths (the ones iv looked at) leading to here:
www.img.discogs.com
and i have seen other sites similar - usually ones that store lots of images (which i am intending to do).
do they just simply purchase a new domain with the 'img.' within it, or is it an image hosting specialised site or..?
if it is simply purchasing an additional site to store the images, is there any information etc on how to go about setting this up i.e. is a login system required on the storage site, or security considerations etc... can image uploading be done through the 'main' site or would it need to be done through the 'img.' site.
i have tried to google, but im pretty poor at naming things correctly and so havn't found any answers as yet.
if anybody could shed some light on this i would be very much grateful! thanks in advance...
why do some sites have a 'img.' site?
wasn't sure what tag to put for this one, if incorrect please let me know
2 reasons.
1) caching policies are usually per subdomain which allows more aggressive caching on IMG. while using normal caching for the main website as it may be dynamic.
2) it may be a different server, a high storage one. The main website is running the main apps hogging all the performance and the image subdomain allows for a separate, high speed delivery.
This question already has answers here:
How do search engines deal with AngularJS applications?
(15 answers)
Closed 6 years ago.
I have a website containing custom elements (i use angular 2), and google fails to parse them correctly :
It only sees
<my-app></my-app>
It seems that the value of this component is not retrieved at all by google robots.
Is there a best practice / workaround ?
Thanks for your help.
This has been asked a million times. Please refer to this question and this specific page by google.
Times have changed. Today, as long as you're not blocking Googlebot from crawling your JavaScript or CSS files, we are generally able to render and understand your web pages like modern browsers. To reflect this improvement, we recently updated our technical Webmaster Guidelines to recommend against disallowing Googlebot from crawling your site's CSS or JS files.
What you can do is getting the HTML of your rendered page and inserting it into the <my-app></my-app> tags without user information or the like.
This will get replaced anyway after Angular has booted up, this means you can even put something completely different in there.
This question already has answers here:
Closed 10 years ago.
Possible Duplicate:
How to prevent downloading images and video files from my website?
I have an HTML page which resides locally on my machine. I have 2 items on every page 1 - PPT presentation 2 - Video tutorial . I want to prevent users from downloading the content for their personal use. How can this be achieved?
Thanks in Advance !
Maddy
Unfortunately the short answer is that it cannot not be done in a good way, if the content should be available at your website at the same time.
There are solutions where you obfuscate the path to the file when it is sent to the browser, and then use a JavaScript to "decrypt" the path at the client. But those solutions are in no way bullet proof, as the decryption technique would have to be sent to the client as well.
This one is not easy to do, especially if you show them on your page already. As you do so they'll get downloaded on the browser an there's no way, at least i'm aware of you could do this with ease!
All you can do is to make it harder... but it still always be possibile! Even if you could stop them from downloading the file, you can't stop them from hooking a VCR to their video card and re-recording it. Even if you use some protected-path technology to stop that, you can't stop them from pointing a camcorder at the screen.
Same applies to the ppt presentations... as they can view them... users could take screenshots or do whatever to create their own copy!
This question already has answers here:
Closed 12 years ago.
Possible Duplicate:
How does Facebook detect images when adding a link?
hi all, curious how facebook is able to pull an accurate image when you post a link. I know they first look for a meta tag with the image the site owner wants them to use but if that doesn't exist what logic do you think they do to get what image should be presented if you link to a blog article let's say.
They most likely grab the source of the page and analyze it to extract the biggest images. They probably look for images close to the center of the page, images that are prominent. It's also possible that they analyze other similar pages on the site to determine unique images that aren't common across pages (this would help FB rule out images used for navigation, headers, decoration, etc.) Not sure how it determines how many perspective images to grab, though.