Behance API does not display all projects? - json

The issue specifically is that when I am trying to retrieve the data from my behance portfolio, I only get 7 projects. I uploaded approximately 30 projects and they are not being returned by the Behance API.
It is not showing all the projects that my account/user HAS. I simply want to get the covers of each project and display them. This works with other users such as Ilya Kuvshinov.
I am using localhost to test the API and only get 5 out of 28 projects.
Does the Behance API take a long time to update information? Is there a setting I need to have to get the API to show all projects?

Behance's API is limited to very low amount of requests (150 per hour) so maybe your problem is one of these two (i could help you out better if you shared some code with us):
Your entire projects need more than 150 calls
You have restricted projects that will not shown to an unauthorized user (like nudity projects with adult content)
To solve first problem you had to manage your API calls under 150 per hour, or simply contact to behance support team and share your problem with them. For second one you most change your API request to an authorized API request that's need to follow behance API documentation. Simple as that, kind regards

Related

Why is the OneNote API lagging by roughly 3 days on my account?

In short, getting page information for my notes through the OneNote API has data that lags by about 3 days. Why is this the case (I assume only for me) and how can this be fixed?
I'm working on a personal R program to read my OneNote Notes. Got it up and running about a week ago. Authentication works fine and I can use GET requests to get information from the OneNote API. However, the information that is returned about my OneNote pages (using GET https://graph.microsoft.com/v1.0/me/onenote/pages) is about 3 days old. New notes from the past 3 days and changes made in the last 3 days are not reflected in this data. It is not frozen in time, i.e. checking tomorrow will give me information that is 3 days behind tomorrow. This is not solely an issue with my application, as the same behavior/results are produced through the Microsoft Graph Explorer. Does anyone have insight on how to fix this?
I can confirm this behavior is still true as of Jan. 30. Calls to get the pages of a section, or pages in general return indexes of very stale information. To test, delete or create a bunch of pages in your OneNote account, then call in the API Explorer:
GET https://graph.microsoft.com/v1.0/me/onenote/pages
You will be given information about deleted pages, and no new information for at least days after these changes have been made. If you follow the returned 'contentUrl' properties for deleted pages, the content does not exist.
It seems other endpoints in the OneNote API are not subject to this same problem. Adding/removing/updating sections, for example, results in immediate changes to the results from the API.
Sorry to hear that. Just to clarify, are you using your personal account or work account for auth?

Use multiple Google Map Account for websites on the same server

I have one real website using google map and It is binded with several domains. Each domain serves for a separated group of client.
Google map announces new policy updates that apply from 11th june 2018 that decrease free limit usage from 25K map loads per day to 28K map load per month
My questions is Can I use map api key which generated from separated google account to increase free usage ?
I have read this question carefully use multiple Google Maps API Keys for same server?
But link seems broken and I can't find relevant information as answer on that question or some other questions too
In my case, although my sites have the same code based and IP, however they works like different sites on shared hosts. One more thing, Google assigns free usage based on account (not google map key) so I think I can use multiple account to increase limit
So can I do this ? I don't want to break Google rules and I appreciate for any comment
You can not since is against Google terms, please check section 3.2.4.e in this link https://cloud.google.com/maps-platform/terms/#16-limitation-of-liability.
No Circumventing Fees. Customer will not circumvent the applicable Fees. For example, Customer will not create multiple billing accounts or Projects to avoid incurring Fees; prevent Google from accurately calculating Customer’s Service usage levels; abuse any free Service quotas; or offer access to the Services under a “time-sharing” or “service bureau” model.
I'm not sure if it's against the rules:
Google Maps guy said here, that it's OK https://twitter.com/GMapsPlatform/status/1001061642363461633

Referrer limit per google-maps api key

We are providing websites/CMS solutions for more than 2500 customers. Almost all websites have google-map module. So since google changed its map usage policy, from one day to another all those webs had an error on their map modules. We need to come up with some quick (and dirty) solution. We decided to use multiple api-keys, and devide domains between them - alphabetic. And we registered all those 2500+ domains under these keys - manually. One by one.
The solution worked until last week. Now we somehow reached some kind of limit, as we cannot register any new domains/referrers under one of those api-keys. The actual count of domains/referrer of this given api-key: 1537. The saving process yields an error with tracking code (which is every time I try different).
Is there really some kind of limit? Does anyone experienced the same problems. Does some time-economic solution exists?
Thanks for any help or suggestions. Peace!
There is indeed a limit of (at time of writing) about 1,000 referer restrictions per API key. You can create about 100 keys per project, so you can authenticate 100,000 domains with a single project. To proceed further, you can create multiple projects (note that multiple projects can be combined under the same billing account, so you would still receive a single bill).
As a short term fix, you can temporarily remove all restrictions on the key, so that apps relying on that key are functional again. Then you can take the time to release a new key sharding pattern that follows these guidelines.
I just created a feature request so that the situation can be improved, for this use case ("star" it, to be notified of updates).
Google has recently released an alpha version of API that allows manage API keys programmatically.
The best way to handle thousands of authorized domains is to use an API to programmatically manage your API Keys and their restrictions, and we have recently launched a new service that allows you to do this.
This API is still in Alpha. If you are interested in becoming a Trusted Tester for this service, you can use the following form to sign up, please read the instructions carefully:
https://forms.gle/qx2SMcarWCAsbWVp7
Please note that this API is not part of the Google Maps Platform. After you fill out the form, you will be contacted by the API Keys API team with instructions on how to get started, and how to receive support.
API Keys API is currently free of charge. However, please note that use of Cloud Endpoints may be subject to charges at high traffic volume. You can check the pricing sheet here:
https://cloud.google.com/endpoints/pricing-and-quotas
source: https://issuetracker.google.com/issues/35829646#comment12
Hope this helps!

Get a users Xbox Achievements list?

Is there a way to get a user's achievements list from Xbox without dealing with screen scraping etc.?
How do sites like http://www.trueachievements.com do it? (they're definitely not screen scraping as they do not have my Live ID credentials...)
I know there's the Xbox LIVE Community Developer Program (XCDP) that provides some API but they do not provide achievements information as far as I can tell (links below)
http://duncanmackenzie.net/Blog/put-up-a-rest-api-for-xbox-gamertag-data
http://file-pasta.com/file/XML_Feeds.pdf
If you want this information you can check out my new site https://xboxapi.com.
Again if your profile is public then you can see friends, games and achievements. This has taken me over a year (on and off) to code and is now in public beta. To date it has served nearly 2,000 requests and growing. I am still working on optimising this script and it should only get better in time.
I currently have an API request limit setup for 120 requests per hour per IP.
Lets hope there are no major updates to xbox.com any time soon.
It says at the bottom of the TrueAchievements website that they are a member of the Xbox LIVE Community Developer Program. That leads me to believe that the XCDP does give information about achievements. Duncan Mackenzie's API for xbox worked for me for a little while, but then his web service seemed to go offline. Duncan also says that his web service will not be returning sadly. I can't get ahold of him to ask for release of his source code. I'm currently looking for another solution in retrieving the data. Unless I can get official access with XCDP, it looks like I'll have to do screen scraping. The screen scraping approach is not going to be pretty though. I'll let you know if I get any more information.
The current XCDP API does not give you any individual achievement information. At the moment the only way to get achievements is by page scraping and as of a couple months ago Microsoft has removed their non-JS log in form for xbox.com making it very difficult to log in. TA has found a way to get around this but he's trying to monopolize the scene and isn't willing to help others on this topic.
Even though TA does not have your LIVE credentials if your xbox.com page is public and if you give them your gamertag anyone can get your information by going to http://live.xbox.com/en-US/GameCenter?compareTo=GAMERTAGHERE
With the change the Microsoft has done they've also modified achievement icons. If you're looking at a friends profile and if you don't have the game or have unlocked the achievement the icon will be monochromatic even if it's unlocked by that person.

Integrating CRM with Google maps

Just started testing Zoho Crm as a CRM solution for our company. Someone asked for a Google map on the page showing our upcoming engagements.I know Zoho provides an API that allows accessing its data from the outside, but I actually need to integrate the map on the data-entry form.If anyone could provide a pointer to any mashup with Zoho CRM (be it Google MAps, Bing Maps, or any similar web service), I would be extremely grateful.
I know this is an ancient question, but since there's no answers and this is pretty much all that came up on google when searching for Zoho CRM integration with Google Maps I'll take a stab at this anyway. I recently got a similiar request, but in this case they wanted to display the leads on a page outside of Zoho.
I created a Java servlet and JSP that runs on Google App Engine. The servlet will connect to Zoho CRM to retrieve all leads and geocode the addresses they are registered with. The client-side Javascript is then taking care of creating the markers on the map for all the addresses.
It's a bit too much code to paste here (although not that much), but you can check it out at http://code.google.com/p/zohomap/.
I put the demo up at http://zohomap.appspot.com/.
I know this is an old question, but it came up on Google Search. About three years ago, I start a similar Google Maps integration project for SugarCRM. The JJWDesign Google Maps project is up on GitHub.com. The idea came about during a marketing meeting and quickly grew out of control.
Download at:
https://github.com/jjwdesign/JJWDesign-Google-Maps
Here are some of the pitfalls that I've experienced:
Exceeding Limits of Geocoding: The Google Maps API v3 has in place a limit of 2,500 Geocoding requests per day. It is also throttled to 10 per second. So, you'll most likely need to develop something to queue these requests. I used a CRON/Scheduled Task to handle the processing trigger.
PHP Memory Limits: The design of SugarCRM creates rather large objects for each one of it's records. Using 10,000 of these objects will usually exceed the memory allowed for PHP to execute. So, special consideration may be needed in examining the best way to pull data into the map.
Always develop/test with a large data set; 10,000+ records. This way you'll be able to more easily see inefficiencies in your code; especially JavaScript. The IE Browser has been know to cause issues with MarkerClustering.
Get ready for an explosion of interest in advanced search / filtering functionality. Also, expect to develop a large section of Admin configuration. Everyone wants something slightly different.
Cheers,
Jeff