Hi Stack Overflow community,
I'm a bit of a noob here (please be gentle) and wanted to ask how to upload HTML/CSS + Packages in bulk to my site.
I'm familiar with the code injection/CSS editor within Squarespace, but something doesn't seem to be working.
To summarize, I received a bunch of files and was requested to upload them to the website I manage. These files contain:
HTML (by page)
CSS (by page)
image files
index.html
Scripts file (which include .js & .php
sitemap.xml
That being said, I know there is a lot of referencing between these files and wanted to know the best route in incorporating these into my site?
Thanks so much!
A quick way to upload multiple files is to use the (S)FTP protocol. You can use an FTP client such as FileZilla to upload files in bulk to your server.
However, I'm not quite sure if that's possible in your use case. Are you using Squarespace for your project? If so, it looks like you can use either Git or SFTP for file uploads. You'll need to have developer mode enabled for that, though.
I found this article that goes into uploading multiple files to Squarespace via their own upload system, does that help?
Related
Summary:
I am a beginner to HTML and need to work out the URLs (or how to make them) of files I have added to 000webhost. I have been given the URL of my index file, and can access it easily, but the links I have placed in it do not work, as I cannot find/don't have the URL of the links (though I do have the code). Is there a way of finding/making a URL for each of the files I have added to my 000webhost project?
So I'm a beginner to HTML and after making a basic website (including links to other pages I have created)I have decided to try and upload it to the internet. I watched a couple YouTube videos on how I should do so and ended up using htmlsave.net . I copy and pasted my code in, changing all my links from places om my desktop to the URLs provided by the website, and everything worked. However, since I had not paid money for a membership I quickly reached my limit on how many pages I could reach. Because of this, I decided to use another(free) web host that would not limit me on how many web pages I could add to a website.
After some research I settled on 000webhost. Everything started off smoothly, I created the project, added my files and got my index file up and running. However, from my index page (which I could now access on the internet), I could not use the links inside it, as they were still still linked to locations on my computer.
Therefore, I opened the code to edit it, but then quickly realized I did not know the new URL of all the files(excluding the index file (named index.html) which 000webhost had provided me as stated earlier) I had added to my 000webhost project.
So after looking around on google and stack overflow I have not been able to find a solution on how to find out the new URL given to the files I have added to 000webhost.
(Apologies for any incorrect use of terminology, as stated I am extremely new to HTML)
I was trying to export a Google Site I made for a project. I used wget to spider through every page and to download the html files and linked content. When I try to open "index.html" in Chrome, it does open the local HTML file, but it redirects me to the live version immediately after.
Is there anyway I could modify the HTML code so that it won't head straight to the actual website? I just want to have a local copy of it for reference, and I don't want to store it on Drive.
As the HTML file is too big to type out, I have provided it on Pastebin here.
.
You need a better question. No website works offline, or they do if you download all the files to your user’s computer so the user can view it offline. But at some point they had to visit it online to get it.
Or you save it as an html site and hand it to them on a USB drive. That’s offline to that extent. But then it’s not really a website, its an html file.
Or otherwise, if you need a website for your school which can be used by anyone through internet / intranet, you have two options -
1. Create and host a website in an online server
. a. You have to buy space and deploy a server yourself.
. b. They will a run website in their webserver for you. You just need to give money
2. Deploy a webserver in the school's any one machine and get it in other machines.
Rephrase the question for a better answer.
My knowledge about Web technologies is very low and I just wanted to know if the following scenario would be possible with HTML5 and Javascript:
If I host an HTML file in Dropbox and send this link to seomeone, would it be possible that this HTML file creates a new file in my Dropbox? For exampe the HTML file is a form that one can fill out, can the HTML file create a text (.txt) file with the form content?
As far as I understand, the HTML file has to be hosted by a webserver and has to allow Javascript or PHP to achieve this. But maybe there is a way to just use an HTML file, a dropbox and a browser?
Any hints what topics I should study to achieve this goal?
On what I've understood from Dropbox, it does not directly show you the file contents in any manner. You can store files there, but the only thing you can see when opening a link that directs to the file, is the page which allows you to download the file to your own PC and save it.
This would seem like an impossible thing to achieve, in any cloud service like Dropbox it would seem. I would recommend you to just get the web hosting service, they are usually not that highly priced after all.
You could do this, but you shouldn't. To make this work, you'd have to use the Dropbox API to upload files, and you'd have to embed in your web page an access token for your account. That means anyone who looked at the source of your web page could get access to make changes in your account (e.g. delete all your files). So there's no safe way to do this without a server-side component (like PHP).
As we all know, the infopath forms service residing on a sharepoint server generates a web site each time we publish an inforpath form template to the sharepoint server.
Here is the question: how does sharepoint do that. Is there any way for us to do that programmatically via some kind of api provided by MS?
In fact, what I need to do is getting all the html, js, css etc. files and applying some kind of operations like deleting some divs or insert some html code into the particular web page. I have come up with two ways to do this.
Generating the web page via sharepoint api and apply those operations at the same time
Extracting the web page files from the IIS server and apply those operations
I am totally new to this kind of work. All in my mind is that each time we right click on a web page in the browser and choose to save the web page, the browser gets some of the files we need to render the web page and makes it possible for us to browse the web page offline.
httrack
WinWSD
and tools like that seems to work fine with extracting html files from online web pages but not that well with js, css files.
Now I am trying to dig into the chromium project for some kind of inspiration, although whether it helps or not is unpredictable.
Any kind of advice will be appreciated.
Best Regards,
Jordan
Infopath xsn files are just zip files with a different extension. you can rename the extension to .zip and extract out the files. you will find a number of files that make up the form. the two main ones are the .xml and .xsl files. the .xsl will have the html to generate when applied to the xml.
I need to add two lines of html to every page on a large website. I will be using a regex to do it.
I would like to know the best tools to download all the html files from a website, then ftp upload them back up on Windows.
Use any standard FTP program and limit transfer to files with a .html extension. Then mirror the files back up.
Or you could just download everything via FTP, make changes to only the .html extensioned files, and then upload modified files back up (should be only .html) files.
I find FileZilla a very decent cross platform FTP client.