App42 cloud service (baas) - bulk upload possible? - app42

Is this possible? Or do I have to upload one document after another? Which needs very long if you want to upload thousends of documents... As with downloading, bulking together 100 documents would be quite nice...
Is this somehow possible?

Through dashboard its not possible however you can write a custom script in which you can pic the files & make a chunk of 50 or 100 to upload your files on App42 server using App42 File Upload API.
Naresh

Related

Upload Bulk Images in Box

We are uploading thousands of file to a Box location thorough Java API. It is taking more than a day to upload these many files to Box. Is there a way to upload files in Bulk in Box using Java API? Would that help in improving performance?
Here is the reference code that we are using: https://github.com/kendomen/BoxJavaJWTExamples/blob/master/src/com/nike/box/UploadFileAsAppUser.java

Monitor and automatically upload local files to Google Cloud Bucket

My goal is to make a website (hosted on Google's App Engine through a Bucket) that includes a upload button much like
<p>Directory: <input type="file" webkitdirectory mozdirectory /></p>
that prompts users to select a main directory.
The main directory will first generate a subfolder and have discrete files being written every few seconds, up to ~4000 per subfolder, upon which the machine software will create another subfolder and continue, and so on.
I want Google Bucket to automatically create a Bucket folder based on metadata (e.g. user login ID and time) in the background, and the website should monitor the main directory and subfolders, and automatically upload every file, sequentially from the time they are finished being written locally, into the Cloud Bucket folder. Each 'session' is expected to run for ~2-5 days.
Creating separate Cloud folders is meant to separate user data in case of multiple parallel users.
Does anyone know how this can be achieved? Would be good if there's sample code to adapt into existing HTML.
Thanks in advance!
As per #JohnHanely, this is not really feasible using a application. I also do not understand the use case entirely but I can provide some insight into monitoring Cloud Buckets.
GCP provides Cloud Functions:
Respond to change notifications emerging from Google Cloud Storage. These notifications can be configured to trigger in response to various events inside a bucket—object creation, deletion, archiving and metadata updates.
The Cloud Storage Triggers will help you avoid having to monitor the buckets yourself and can instead leave that to GCF.
Maybe you could expand on what you are trying to achieve with that many folders? Are you trying to create ~4,000 sub-folders per user? There may be a better path forward should we know more about the intended use of the data? Is seems you want hold data and perhaps a DB is better suited?
- Application
|--Accounts
|---- User1
|-------Metadata
|----User2
|------Meatadata

Partially editing a document in Google Drive

This is a conceptual question regarding the realtime editing of large files. Imagine we have a 50 Mb txt file in Google Drive that we want to allow users to edit. We require that the user downloads the entire file before they start editing (The user will have to wait for a while, but this is ok). The user then changes a single word in the 50 Mb text file. How can we possibly update the file in Google Drive without uploading all 50 Mb of text.
If you are creating an Android application, the Google Drive Android API handles this differential upload for you behind the scenes.
If you are not using the Android API, it is not currently possible via the Drive RESTful API.
You say the files are "txt" which suggests they are not Google docs. Drive considers all such files to be blobs. so I can't see anyway they can be patched.

Is it possible to create a task using the box-api v2 in a folder?

I have successfully created tasks using the box-api v2 on files. However, I often need to create a task in a folder, e.g. for someone to upload a file to the folder, not simply review a file in that folder.
It doesn't appear to be possible now, is that true? if so, any plans to make that possible?
It's not possible to assign a task to a folder via the API or the Box web application. I don't believe Box has any current plans to add tasks either via the web or programmatically.
Rory

How i can get multiple files from google drive through the google drive api?

I would like to know how i could to obtain multiple files from google drive. I searched this in the reference but i not found this information. I'm building a web application that will talk to drive and retreive a link of a zip file to download. Zip of the files.
I'm using php with api v2.
That is currently not possible with the Drive API, you have to send multiple requests to retrieve multiple files.
I've been faced with a similar problem and while there's currently no way of doing this through Drive (to my knowledge), this is the solution I came up with.
I'm serving up hundreds of thousands of files to clients using a Drive folder as the storage with a custom built front-end built with the Drive API. With that many files, it's ridiculously tedious for users to download files one at a time. So the idea was to let the users select which files they'd like to download and then present them with a zip file containing the files.
To do that, first you'll need to get an array of the Drive files that you want to download whether that's some you generate programmatically or through check-boxes on the front-end. Then you'll loop through that array and grab the 'downloadURL' value for each file and perform a cURL to the url provided. Depending on how many files you're planning on handling per request you can either keep them all in memory or temporarily store them on the disk or in a database. Regardless, once you have all of the files, you can then zip them up using any number of zip libs that are out there. Then just send the resulting zip file to the user.
In our case we ended up sticking with individual file downloads because of the potentially massive amount of resources and bandwidth this can eat but it's a potential solution if you're not serving large numbers of files.
Assuming I am answering the correct query, if you place the files in a folder on google drive, then as far as I know it is possible to download as a complete group of files.