I recently created the first 2 apps of my life and triggered the publication process. I noticed that Microsoft services certifies apps within a day or two now it has been almost 4 days and there is no news of the certification passing or failing. Please tell me how i can get my app certified faster.
Microsoft says it usually takes 16 Hours for Certification. But that is totally not the case. One of my app's Last certification took 2 days (This app had FB, Twitter and Youtube Integration). i updated another app before that with minimum functionality and it took less than 10 Hrs.( this was a very simple app helpful for developers )
Here is the URL that describes Certification process
Related
In short, getting page information for my notes through the OneNote API has data that lags by about 3 days. Why is this the case (I assume only for me) and how can this be fixed?
I'm working on a personal R program to read my OneNote Notes. Got it up and running about a week ago. Authentication works fine and I can use GET requests to get information from the OneNote API. However, the information that is returned about my OneNote pages (using GET https://graph.microsoft.com/v1.0/me/onenote/pages) is about 3 days old. New notes from the past 3 days and changes made in the last 3 days are not reflected in this data. It is not frozen in time, i.e. checking tomorrow will give me information that is 3 days behind tomorrow. This is not solely an issue with my application, as the same behavior/results are produced through the Microsoft Graph Explorer. Does anyone have insight on how to fix this?
I can confirm this behavior is still true as of Jan. 30. Calls to get the pages of a section, or pages in general return indexes of very stale information. To test, delete or create a bunch of pages in your OneNote account, then call in the API Explorer:
GET https://graph.microsoft.com/v1.0/me/onenote/pages
You will be given information about deleted pages, and no new information for at least days after these changes have been made. If you follow the returned 'contentUrl' properties for deleted pages, the content does not exist.
It seems other endpoints in the OneNote API are not subject to this same problem. Adding/removing/updating sections, for example, results in immediate changes to the results from the API.
Sorry to hear that. Just to clarify, are you using your personal account or work account for auth?
I'm considering implementing and publishing a Web App using Google Apps Scripts, which I'm quite familiar with for personal projects. I've never published a project to share with other people but I have an idea that would fit perfectly well for consumer users, probably for thousands of them. I've googled for the last 24hrs to see if there's any experience with simultaneous executions of a Web App when it runs as "user accessing the app" but I've failed to find any direct answer to my questions.
According to Google Developers Documentation, there is a daily limit of 30 simultaneous executions of the Web App, which coincides with the results of this test. However, neither the documentation explains on what conditions that limit would be reached nor the test was run with different users accessing the Web App authenticated.
Before trying to test it myself, which I find hard to do since right now I have no access to more than 30 different Google consumer accounts, I'd like to ask if there is someone who has direct experience publishing Web Apps using GAS with permissions set as "user accessing the app" and has found errors beyond the 30 simultaneous users.
Thank you.
Edited: 14/jul
The issue specifically is that when I am trying to retrieve the data from my behance portfolio, I only get 7 projects. I uploaded approximately 30 projects and they are not being returned by the Behance API.
It is not showing all the projects that my account/user HAS. I simply want to get the covers of each project and display them. This works with other users such as Ilya Kuvshinov.
I am using localhost to test the API and only get 5 out of 28 projects.
Does the Behance API take a long time to update information? Is there a setting I need to have to get the API to show all projects?
Behance's API is limited to very low amount of requests (150 per hour) so maybe your problem is one of these two (i could help you out better if you shared some code with us):
Your entire projects need more than 150 calls
You have restricted projects that will not shown to an unauthorized user (like nudity projects with adult content)
To solve first problem you had to manage your API calls under 150 per hour, or simply contact to behance support team and share your problem with them. For second one you most change your API request to an authorized API request that's need to follow behance API documentation. Simple as that, kind regards
I trained a new English to French MT. Training went ok with a 62+ BLEU score, but deployment is stuck. Any idea what's going on? I've been able to train and deploy in the past. This is the first time since the move to Azure. Anything new I need to know?
This is the screen I get. It's been pending since 07/13.
Here is more details.
When I click on Request Deployment, I get this message:
Your request to deploy this system was submitted on 13-Jul-2017 by XXXX.
Deployment will take up to 2 business days. Project owners will be notified when the system is deployed.
Deployments failed for all customers since Thursday July 13. Expected to resume fully on Tuesday July 18.
Customizing NMT is now available using Custom Translator (Preview). The new Translator Text API v3 works with Custom Translator, and the supported languages are limited to pairs where NMT languages exist.
Thank you.
I have written an entire application to do one and only one thing, notify the people on a shared folder in drive that a new file has been added to the folder. the fact i had to write the application is baffling to me but whatever.
The person generating the content clicks on the notify button when he adds a new file, which on a good day would be about 2 or 3 hundred times, the system will go out, find the new files, get the list of people its shared with based on the folder, and generate an email for those people saying their file is ready to download via the link in the email.
my problem is he starts work about 8 in the morning and about 9 in the morning he gets a notification that he has hit the email quota for the day.
so, just to save time let me answer some of the mundane questions
he has a google apps for business account
no, he can not simply click the notify button less each day. he typically works on
a 5 to 15 minute turn around time for the files
in a given year he averages about 14,000 files added to the system with 200 clients
and i would guess about 600 emails a day.
yes, i have rewritten the app several times now to minimize the number of times the api's are called. I can't go any less than i am currently calling them because the toolset just isn't that good.
the server that ran this system previously was a P4 1 gig processor with 2 gigs of ram
and a 500 gig hard drive that sits in his basement on the end of a
128k ISDN line. that server handled this task well for over 10
years, i think google can handle the load
none of his clients have gmail accounts to be notified of new files as a shared folder
i have his credit card ready to upgrade the quota for the gmailapp api and the driveapp api inside the google app script language and i only need to know what screen to enter it to upgrade these quotas.
Am i going to have to rewrite this app yet again as a GAE app to make this work as expected?
thanks
As an update, it turns out no-one is sure if these limits can be increased. Therefor I rewrote the app in GAE. At least with that framework and Drive API, if he ever hits any of the limits for the free account, he can be back up and running in minutes with a credit card, not have his entire business unceremoniously shut down for 24 hours with App Script.
Thank you all