Trying to understand this help doc. It says:
The Realtime API provides a persistent document for the data that
users are collaborating on.
What is this "persistent document"? Is this a file in Google Drive? I'm not planning to store data in Google Drive and most likely will be using shortcut files. So how does that relate to my situation?
Applications that provide collaborative
editing features, but need to work with files in various formats, must
adopt one or more strategies for converting data between the Realtime
API document and other file formats.
I'm storing my data as XML in database. Does this qualify for "files in various formats"? How does that effect my situation?
Each file on Drive has content as itself and a permanent storage for Realtime API. In your case, your shortcut files will have Realtime structures even though your files wont have any content all.
Related
I`m quite new at using saleforce but I have been able to use Cloud Data Fusion recently to transfer information from a Data Extention to a table in Big Query.
Know I would like to do the same thing but with a Shared Data Extension, meaning that (possibly) features like the external key of the D.E would not exist because the D.E is shared by different Business Units (B.U) and also, I think I would have a problem by dealing with the authentication (Client ID – Client Secret - ) because again, is a Shared Data Extension.
Thank you very much for any insight on this matter.
How can we maintain multiple versions of the same DWG or other 2d file while uploading and translating the file. What are the preferred options.
I need to maintain the versioning of the files in Private OSS Bucket with Autodesk Forge and not using BIM360 in this case.
Thanks a lot.
OSS is just a simple blob storage service and as such it does not provide any versioning capability. You can introduce your own versioning schema, though, for example by appending a version suffix to the object key of every uploaded file (and then removing the suffixes when listing the objects), or by maintaining the document-version relationships (and then perhaps even folder-document relationships) in a relational database, with version records pointing to OSS objects.
I have about 2 million text searchable PDF files. I need to be able to search them for a user's query and return a snippet and a filename. The front-end is a Node.js React app.
Right now I'm able to use pdfjs-dist (https://github.com/mozilla/pdfjs-dist) to read the PDF content into a MySQL db. Then use fulltext MATCH ... AGAINST queries to search through the text. However, this is awkward and with 2 million PDFs is really slow. Also, new files are added regularly, so reading the PDFs into SQL is pretty resource intensive, too.
Is there a better solution? Is Elasticsearch a good solution for this?
The project is hosted on Google Cloud (App Engine & Cloud SQL). Is there a Google tool that can do this?
Yes, I would say that Elasticsearch is a great tool to index PDF and search in it later.
There is an ingest attachment processor plugin that allows to extract data from the common formats (PDF, TXT, DOC, etc) and index it into Elasticsearch, so later it could be searchable.
Google Cloud have Elasticsearch cluster plugin which should make integration easier. Also there is a Elasticsearch service in the Google Cloud which have massive support.
We are planning to implement a feature to let customers browse their images in their own google drive or dropbox accounts (within our app), and select ones they wanna use in our app. Then, we will save these images into our Google cloud storage (GCS) place.
A simple approach is to use OAuth to download the customers' images from dropbox/drive to our local disk, then upload them to our GCS.
I am wondering if it is possible to make a direct transfer between dropbox/drive and GCS? GCS does provide transfer service for S3. You can also create a list of URL list, but each URL needs to be publicly accessible, which in our case does not seem possible.
Is there any better suggestion on how to implement this feature? Thanks
Well you could stream the files so at least you wouldn't have to save them to local disk, but other than that I think your solution is the way to go.
It depends on your language how streaming works exactly, but basically you download the file and upload it to GCS right away without ever writing any bytes to local disk.
I ran into your question during my search of finding a way to transfer data from G-drive to Google Cloud Storage. Since they are both under Google umbrella, so I guess they should be able to seamlessly upload/download. So far in my search, without programming, the answer is NO.
I don't have a exactly solution for your question, because it really depend on how much data you have and how often of the uploading/downloading. Unfortunately, there is not straight shot (even for the Gdrive!). But maybe thinking of trying the RESTful API calling to Google and make the connection with Dropbox API. Hoping this could help.
With this recipe, you can use Google Colab to mount Google Drive folder, and then copy from the mounted folder to the storage bucket using gsutil, avoiding the need to first save the data to local disk.
Can gDrive be used to backup application-specific data files, i.e. game save states that are not traditionally user facing? Would this allow cross-device migration?
Yes, this is possible. Drive apps can create files with arbitrary MIME types. Synchronization is therefore available by downloading the file(s) from multiple places/devices.