UI Path Studio - Data Scraping - Cannot find pattern - palantir-foundry

This website (Palantir Foundry - which is a pay-only cloud service, so it is not publicly available) Seems to use some React Tables package that creates a table with headers (that one can sort, filter, etc) and then rows of data.
We want to capture the header row and the first data row with UIPath Studio (Enterprise) 2020.10.2
When I can use the Chrome extension to grab a row, but when go for the second element in the pattern i get the pop-up error “Cannot find all pattern elements”. Any suggestions on things to try?
Pictures below might help:

Related

UI Testing (casperjs) with good known status of data (mysql database)

I'm using CasperJS for automated UI tests. I've done the basic UI testing and validation with some random data, kind of POC. I've set up this automation using bash script which kicks to start the web server, load MySQL data from SQL file, start CasperJS test cases, stop the web server, check the log files.
Now, I want to start the testing with some good known status of data which are stored in MySQL. So that I can test the list data and form data with detailed field information with some known database status. How should I know the status of data in the database at a moment?
1) Should I use pre-populated JSON dumped file which has status and details about all data?
2) Should I use web service API? (web service APIs are being used to show/save/delete data from the web page)
Let's take an example. I've 5 users in Users table. Now when I open the home page it shows 5 users with some rough details. When I click on any record from the list of users, it shows a form with detailed information about that user. The webpage is requesting to the web application to get the detail about a user with the help of user_id to show the detailed user data in a form. Now I want to check that all the data in that form is populated correctly. So at the next step, what would be the preferred way, should I read content from JSON dumped file or should I use web service API (like webpage does).
Searching this problem online, I also found MYSQL HTTP plugin. Should I consider this as well? and How safe it is to use? (I know from the docs that this plugin is not for the production, it is just for testing purpose only. :) )
For the main question in cases like this I would change the database connection string to your testing database (this is a clone).
In your case use your bash script to change the connection string (file copy?) automatically before you run the tests. And when completed change back.
Your testing database is a direct clone of your dev/live databsae but with ONLY the test data you want. Downside is you need to keep the schema in sync with DEV/LIVE.
Also another point to take into considertion is if your testing changes state (post). If so your testing data might be out of sync. One way is get around this is to drop foreign keys, truncate the data and load in a dump file.
HTH

Google Data Studio Gives Error When a 3rd Schema Attribute is Added

I have created a Custom Community Connector that currently has a schema with two attributes. A Metric: Budget Amount, and a Dimension: Budget Name. These are pulled from an API from another site. When linking this connector to a Data Studio project I am able to create a Pie Chart that displays all Budget names and amounts.
However, once I add a third attribute of any kind (ex. Metric: Budget Spent or Dimension: Company Name) the pie chart no longer shows and displays an error :
Script error message:
Script error cause: UNKNOWN
Script error stacktrace:
I have confirmed that data is flowing correctly by using logs within Apps Scripts and with the Table chart in Data Studio. The Table Chart is the only table that will show up but only when all possible Metrics and Dimensions are apart of it.
I am currently looking for anyway to debug this, solutions, or advice on where to go from here. I would like to expand the Schema, however I cannot continue if charts stop working when I add more attributes.
If more information or code is needed let me know and I will provide whatever is needed. Thank you for your time
The Debugging Guide should get you started with the debug. I recommend you use Stackdriver logging to log the request and response for each function. You can see an example of logging here.

Inconsistency in MS Graph API behaviour for onenote

When a section renamed get sections API doesn't reflect the updated name whereas get page api shows updated parent section name. This seems to be bug/ data inconsistency in ON API.
On change of anything at page level updates the lastModifiedDateTime for a section but nothing gets changed at notebook level. This again seems to be like some data inconsitency issue.
Can somebody clear this confusion.
(Note - All above can be tested using MS Graph API Explorer
)
These are two separate topics:
Section renaming
This is a known limitation/bug in OneNote - if you rename a section in OneNote Online (in your browser), then the API GET ~/notebooks/id/sections or GET ~/sections will give you the "old" name. This is because OneNote Online doesn't actually rename a file, it only marks the file as "to be renamed" - if you were to look at the file itself in OneDrive/SharePoint it would still have the old name.
Once the OneNote Native Client sees the section (for example OneNote for Windows) sees the section that has been marked as "to be renamed", it actually renames the file.
The OneNote API GET ~/sections/id/pages actually looks at the section binaries and is able to tell whether the section is renamed or not, which is why that name can be trusted as the "most up to date" one.
I have communicated this feedback to our team and we are exploring alternatives - I encourage you to start an item in uservoice so we can better understand impact.
https://onenote.uservoice.com/forums/245490-onenote-developer-apis
LastModifiedTime (LMT) on notebook/section clarifications:
The LMT of a section is equal to max(LMT of pages under it).
The LMT of a section group however is not max (LMT of sections and section groups under it). A section group is a folder and its LMT should behave like that of a folder in a traditional file system (reflects time of last add/delete of a file/folder directly under it).
However, there is nothing stopping you from using $expand and calculating the LMT (as you understand it) yourself based on the entities below the notebook/section group.
https://blogs.msdn.microsoft.com/onenotedev/2014/12/16/beta-get-onenote-entities-in-one-roundtrip-using-expand/

No Google BigQuery table created after importing data through webclient

I'm currently familiarizing myself with Google BigQuery by working through the examples at https://cloud.google.com/bigquery/web-ui-quickstart. Doing a query over the pubic datasets runs fine.
I run into problems when uploading custom data into a new table through the WebUI. I create a new dataset and table, and upload the csv file provided with the example case. As in the example I input the schema and submit the file. Now the upload window stays on top and turns grey as if it's working. Nothing seems to happen afterwards though. When clicking away the upload window after a long wait, the table seems to be created in the tree on the left. However, when clicking on the table an error is shown:
"Unable to find table: ndwtest-984:csvtest.csvdata"
This seems like a trivial action, however I cannot seems to get it to work. I've tried varies different files, uploaded the file to Google Cloud Storage first and played around with the advanced options the last two days, but keep getting the same error.
Help would be much appreciated.
Some steps to help you:
billing must be enabled
you need to choose to upload one single TXT file from the example eg: yob2013.txt and not the zip file
make sure the schema is entered as text: name:string,gender:string,count:integer
on the last screen of the wizard you don't need to change the default CSV option parameters (for demo purposes works as it is)
I just tried the example, and it does work for me. In case you still have errors, than you can check your Job History menu in the Web UI, direct link would be, warning you need to put your Id in the link.
https://bigquery.cloud.google.com/jobs/YOUR_ANONYMOUS_PROJECT_ID_HERE?pli=1

Do Removed / Deleted WordPress Widgets Are Lost Forever?

If a person has some important text on a widget and accidentally he removes/deletes the widget by expanding it and clicking delete in wp-admin widget's area. Then do this mean that the widget has been lost forever?
Do widget resides somewhere in any of the mysql database's table? Are there any revisions of widgets available?
Another way to get the text is to search the web caching engines like Google Cache, Wayback Machine etc. But what if we are unlucky enough that we have not allowed web crawler in robots.txt file?
Update (August 21, 2013) :-
Besides you give a general answer. Also state whether I can recover a simple 'text/html widget' that comes along with WordPress? Thanks.
It depends on the widget:
the widget developer decides where and how to store the widget data and if that data has revisions or not.
At this point you have the fallowing options:
1) access your database and have a look in the "_options" table for for the option_name %LIKE% widget -> it will show you all the data you have saved related to that widget, if you don't find what you are looking for:
2) check the php code of the widget to see if you can find where it saves the data and if it has revisions.
Your best luck is to have the data in the "_options" table and if you have backup of the table you can extract the data from there.