How to Access NYC ACRIS Real Property Master via SODA2 - socrata

How do we access NYC ACRIS Real Property Master via SODA2?
To recreate the issue:
go to: (step 1)
https://data.cityofnewyork.us/City-Government/ACRIS-Real-Property-Master/bnx9-e6tj
Navigate to Export -> SODA API - API Docs
you will end up here: (step 2)
https://dev.socrata.com/foundry/#/data.cityofnewyork.us/bnx9-e6tj
We see this authorization screen and even after allow'ing the same screen appears again and again, it seems to be in a loop. Can you help? This works for tiny files but not the ACRIS Master File, it appears.
Here's a picture of step 2 above:

Unfortunately the ACRIS Real Property Master filtered view is a bit of a weird case that the City of New York has set up for their data.
That filtered view (denoted by the blue "funnel" icon you see at the upper left) is actually a limited access version of a dataset that has been made private, hence the page asking you to authenticate before viewing the API docs for the base dataset.
You can make a limited class of queries through the API endpoint for the filtered view itself:
https://data.cityofnewyork.us/resource/bnx9-e6tj.json
What are you looking to do?

Related

storage inspector firefox local storage edit JSON error

I am working with an online game that allows editing of its code for testers and / or customers who want to mess around on a private level. The game is called Choice of Game: Rebels Uprising. It is a text based game where you build a character and the story/character change based on your decisions. I am trying to test different things and I am having trouble editing the stats which would allow me to easily control testing different options of play. I got this method from their approved forums and have used it on multiple games. However this one has been very tricky and refuses to work.
Once I open Storage Inspector, I normally can go right to local storage and select the data to change, and then refresh the page to enter the changes.
For this game, I have to go to session storage, select the only option there, then go back to local storage which then reveals the data I wish to change. Otherwise it is invisible. The following pops up and what I would like to change:
{"version":"12169","stats":{"fname":"Balin","lname":"de Brackii","orig_lname":"de Brackii","epithet":"","liebreaker":false,"openhanded":false,"eclect":false,"slur":"","shameveined":false,"filthborn":false,"childkiller":false,"alias":"","crest":"Great Elder Tree","wisard":"Theurge","wisardry":"Theurgy","wisardric":"Theurgic","aristo":true,"helot":false,"female":false,"male":true,"femalepref":true,"malepref":false,"acepref":false,"aropref":false,"kuria":"Captain","milady":"Captain","he":"he","him":"him","his":"his","hers":"his","woman":"man","girl":"boy","sister":"brother","daughter":"son","ruth":32,"skep":24,"natl":46,"ruthreal":32,"skepreal":24,"natlreal":54,"cha":"1","com":"0","int":"2","wealth":35,"loot":"0","followers":336,"kids":178,"nonkids":158,"dead":"0","kidsdead":"0","sa_foll":"12","sc_foll":"21","needfood":88,"foodstore":10,"mules":"4","sickmules":0,"freemules":4,"arms":10,"blood":4,"usedblood":"0","theurgy":"1","theu_spread":"0","theuknown":"0","religion":"0","literate":"10"
I tested by simply editing mules and freemules from 4 to 34. I then tried exiting tab and reopening and it, as well as just refreshing the page and both times I get this error:
JSON.parse: end of data after property value in object at line 1 column 1003 of the JSON data
I have even tried editing the temp one as well before closing/refreshing and this did the same thing.
Ideally, I would like to be able to change the wealth, the mules, the arms, and data like that. Please advise on how to correct/avoid this error. Thank you!
https://www.choiceofgames.com/rebels/#utm_medium=web&utm_source=ourgames

Inconsistency in MS Graph API behaviour for onenote

When a section renamed get sections API doesn't reflect the updated name whereas get page api shows updated parent section name. This seems to be bug/ data inconsistency in ON API.
On change of anything at page level updates the lastModifiedDateTime for a section but nothing gets changed at notebook level. This again seems to be like some data inconsitency issue.
Can somebody clear this confusion.
(Note - All above can be tested using MS Graph API Explorer
)
These are two separate topics:
Section renaming
This is a known limitation/bug in OneNote - if you rename a section in OneNote Online (in your browser), then the API GET ~/notebooks/id/sections or GET ~/sections will give you the "old" name. This is because OneNote Online doesn't actually rename a file, it only marks the file as "to be renamed" - if you were to look at the file itself in OneDrive/SharePoint it would still have the old name.
Once the OneNote Native Client sees the section (for example OneNote for Windows) sees the section that has been marked as "to be renamed", it actually renames the file.
The OneNote API GET ~/sections/id/pages actually looks at the section binaries and is able to tell whether the section is renamed or not, which is why that name can be trusted as the "most up to date" one.
I have communicated this feedback to our team and we are exploring alternatives - I encourage you to start an item in uservoice so we can better understand impact.
https://onenote.uservoice.com/forums/245490-onenote-developer-apis
LastModifiedTime (LMT) on notebook/section clarifications:
The LMT of a section is equal to max(LMT of pages under it).
The LMT of a section group however is not max (LMT of sections and section groups under it). A section group is a folder and its LMT should behave like that of a folder in a traditional file system (reflects time of last add/delete of a file/folder directly under it).
However, there is nothing stopping you from using $expand and calculating the LMT (as you understand it) yourself based on the entities below the notebook/section group.
https://blogs.msdn.microsoft.com/onenotedev/2014/12/16/beta-get-onenote-entities-in-one-roundtrip-using-expand/

Passing commands with URL leaves parameters vulnerable

We are currently using a generic report which will be used differently by multiple usergroups. We have made this possible by creating linked reports with different settings of hidden parameters (such as 'show column x', 'enable feature y').
These settings (parameters) are needed on other reports as well, so we pass them along using the Go to ... Action.
To create the look and feel we are after, we are passing some additional parameters as well, HTML Viewer commands and Report Server commands such as &rc:Parameters=False (reference).
Unfortunately, this leaves us with only the option Go to URL, since Microsoft hasn't implemented these commands for Go to Report. This means we have to pass our settings (the hidden parameters) along in the URL. This results in a security issue, example given: &PARAMETER_ENABLE_FEATURE_Y=False.
The user might notice this parameter in the URL and is so given the possibility to enable this function by editing the URL to &PARAMETER_ENABLE_FEATURE_Y=True.
So my question is: how to use an Action in Reporting Services while preventing users from editing our sensitive parameters and while being able to use HTML Viewer commands and Report Server commands?
You will never get complete security in this sense if you absolutely have to use URL based parameters.
When navigating via the URL, the only way you can hide parameter values without hard coding them is to make them data driven. In your scenario however this will not be 100% secure as you will still need to pass the value that populates your data driven parameters.
This level of obfuscation is probably enough and can be achieved by collating a list of either every parameter combination or just the ones your need and assigning it an ID that you can call in a dataset. This can obviously still be changed by your users should they get curious and can be a faff to maintain.
I would say your only other option is to hide the URL bar completely by providing a 'landing page' for your reporting and displaying everything in an iframe. This frame can be targeted with a javascript link in your Go To URL:
="javascript:void(window.open('URL to open','iFrame Name'))"
If you are able to though, I would advise you group your users into Active Directory security groups and then maintain a collection of permissions and customisations per group. You can then check which groups a user is a part of using custom code similar to the answers here and return the required parameter values accordingly.
Doing things this way will also enable you to maintain which groups can see what from a central location, assuming you have rolled out the same parameter structure across all reports.

No Google BigQuery table created after importing data through webclient

I'm currently familiarizing myself with Google BigQuery by working through the examples at https://cloud.google.com/bigquery/web-ui-quickstart. Doing a query over the pubic datasets runs fine.
I run into problems when uploading custom data into a new table through the WebUI. I create a new dataset and table, and upload the csv file provided with the example case. As in the example I input the schema and submit the file. Now the upload window stays on top and turns grey as if it's working. Nothing seems to happen afterwards though. When clicking away the upload window after a long wait, the table seems to be created in the tree on the left. However, when clicking on the table an error is shown:
"Unable to find table: ndwtest-984:csvtest.csvdata"
This seems like a trivial action, however I cannot seems to get it to work. I've tried varies different files, uploaded the file to Google Cloud Storage first and played around with the advanced options the last two days, but keep getting the same error.
Help would be much appreciated.
Some steps to help you:
billing must be enabled
you need to choose to upload one single TXT file from the example eg: yob2013.txt and not the zip file
make sure the schema is entered as text: name:string,gender:string,count:integer
on the last screen of the wizard you don't need to change the default CSV option parameters (for demo purposes works as it is)
I just tried the example, and it does work for me. In case you still have errors, than you can check your Job History menu in the Web UI, direct link would be, warning you need to put your Id in the link.
https://bigquery.cloud.google.com/jobs/YOUR_ANONYMOUS_PROJECT_ID_HERE?pli=1

How do I get the contents of my webbrowser-control to be visible in my MS Access report when printed?

Un-answered duplicate here: MsAccess Webbrowser Control on a Report not working
Hello all. I have been unsuccessful in my attempts to do the following:
To make the contents (in this example, a static google map using the api) of my web-browser control visible when my MS Access 2010 report is in print preview and when I export it as a pdf.
For my working example, I have the following three objects in my Access 2010 database:
1) a table - called "tbl_Locations" (which contains location data on three sites)
The table is as follows:
LocationID | Location_Name | Address | Zip |
1 Site A 162 East 33rd street 10016
2 Site B 550 1st Avenue 10016
3 Site C 151 West 26th Street 10001
2) a form - called "frm_Locations"(the control source of this form is 'tbl_Locations'). Notably, this is the form that contains the webbrowser-control object (where a static google maps is generated from the data in "tbl_Locations").
Form in Design View:
Form in Form View:
3) a report - called "rpt_Locations" (where I have embedded the 'frm_Locations' object)
Report in 'Report' View
Report in 'Print Preview'
If you clicked on the picture links above, you could see that the webbrowser-control, and its contents, are visible in the 'Form' view for 'frm_Locations', and 'Report' view for 'rpt_Locations'.
However, this is not the case in the 'Print Preview' view of the 'rpt_Locations'. Additionally, I am unable to see the webbrowser-control, or the static google map within it, when I export the report object as a pdf.
I would like to know why this happens, and if there is a VBA solution I can use to ameliorate this.
Static Maps API seems to return a jpg of a Google Map from a URL
https://developers.google.com/maps/documentation/staticmaps/
Then you could load the jpg in your browser or embed the jpg in your report and presumably that would appear in you pdf.
Just like this url
https://maps.googleapis.com/maps/api/staticmap?center=40.714728,-73.998672&zoom=12&size=400x400
provides this map
From what I've read on the web, I'm not sure that it is possible to do this.
I've found a couple of posts which imply hope, but neither indicates concrete success. These are here and here
Alternatively, if instead of trying to do it all from access, you create the pdf first, then you may be able to embed the link you require afterward as shown here