Hello I'd like to know if google app script/google forms/google spreadsheet is a good way to achieve the following. (Similar to previous posts on google scripting, i am struggling a bit to understand the 'landscape' of coding problems that these google tools are good for):
A cohort of N players (could be hundreds) is asked to click a link to a web page where they enter a player id
Once the id is entered they see a message that asks them to wait until all the cohort has entered their id, and a scenario and a question (text).
Once everyone has entered their id players are paired at random and can answer the question by simply selecting a number from a drop down menu (e.g. 1-21)
The numbers given as answers are then scored according to a simple payoff matrix and each player is presented with an outcome based on the difference between the numbers they submitted as answers.
The game then repeats i.e. the players are asked the same question, and paired randomly again (they shouldn't need to re-enter their id)
What I want is a simple way to create this workflow/process/game, and a simple way to monitor the progress of the game. Something like a google form for the questions and answers and a google spreadsheet where the data is visible by the researchers.
What I don't know is whether app script and google forms/spreadsheets can be used to setup this process/workflow.
Forms + spreadsheets would also be neat because there are many versions of this basic game/process/workflow in economics i.e. different scoring matrices and questions.
I don't think you can do this in Forms, as it is not all too dynamic (not too sure), but you can easily achieve this with HtmlOutput.
You'll have to setup a linking sheet, it will store everybodys answers, and you'll constantly check this sheet for new values with setInterval to check periodically, and google.script.run to retrieve the info.
Also, this platform is not the fastest or most suitable (I don't know an alternative), but it's totally doable.
Related
I need to build a PowerApp which presents a daily list of tasks to a user such as number of surveys to complete; within which postcode area etc.
They then want to complete a questionnaire; and store submissions in the PowerApps own Dataverse.
My thoughts are;
Ideally i would want to hook up Microsoft Forms (standard not Pro) so the PowerApp could either link to the form and pre-populate a unique ID or similar? they complete the form; and a Logic App would process into the DataVerse by unique ID. However not sure this is possible?
Other option is i design the Canvas App to complete the survey (similar to the Employee Engagement survey template) - but likely will have 30/40 questions - this would obviously have a direct link to the Dataverse so no Logic App required
Other options?
Thanks in advance
Rich
Stay away from Microsoft Forms for a 30-40 question survey. The conditional logic, error handling, input validation, etc are not robust enough. You'd likely get too much free text to produce any meaningful reports from.
Caveat being: Do external users (outside your Active Directory) need to access the survey?
If so, then Forms is likely the only way to go (unless you want to somehow manage PowerApps licenses for external users).
PowerApps will suffice.
I used scripting to create a Google Form with 200 + pages (generally only one question per page, though) and now need to add the branching logic to it. By the way, the new Forms gets really slow with this many pages.
Anyone know if I can use (Google) Sheets together with Forms and a script to create answer-based branching logic? I would envision Column A would be the question numbers and columns B-n would be used to point to question numbers based upon the answer. If needed, I could repeat the answer value in this exercise.
Assuming setChoices along with item.createChoice can do this.
Thanks
I'm new to geocoding so I'm not certain this is even the question I should be asking, but all of the other discussions I've seen on this topic (here and on the Google API forum) are so application specific that I feel like I might be missing a very elementary step - I don't need to know how to implement a store finder - I need to know if I should.
Here is my specific situation - I have been contracted to design an application wherein we will build a database of shops (say, independently owned bars and pubs). This list will continually grow and change as shops close and new ones open. The user can enter his/her point of origin (zip code or address) and be shown a list or map containing all the various shops within a given radius in order of proximity.
I know how to deliver these results from a static database:
One would store the longitude and latitude as columns for each row and then just use that information to check distances.
But I have inherited an (already fairly large) database of shops which have addresses but not coordinates - so I'm not sure what the best way to get those addresses is. I could write a script to query them one at a time against google geocoding, I could have a data entry person manually look up the coordinates for each one and populate the data that way, or maybe there is a third option I'm not aware of.
Is this the right place to be asking this question? Google Maps Geocoding doesn't host a forum of their own, but refers people to Stack Overflow. Other forums on the net dealing with this topic are all relating to a specific technical question but no one seems to be talking about it from a top-down perspective (ie the big picture).
Google imposes a 2,500 queries per day limit on free users and a 100,000 queries a day limit on paid ones - neither of these seem to be up to the task of a site with even moderate traffic if, every time a user makes a request, the entire database (perhaps thousands of shops) are being checked against Google's data. It seems certain we must store the coords locally but even storing them locally, there will have to be checks against Google in order to plot them on a map. If I had a finite number of locations (if, for example, I had six hardware shops) and I wanted to make a store locator, there would be a wealth of discussions, tutorials, and stack overflow questions available to point the way for me, but I'm dealing with a potentially vast number of records and not sure how to proceed or where to begin.
Any advice would be welcome - Additionally, if this is not the best place to be asking this question, a helpful response would be to indicate a better place to post it. I've searched for three days but haven't found what looks like a good resource for asking such subjective questions.
The best way of course would be when you use a geocoding-service to get coordinates and store the coordinates in your DB. But it's not possible with google's geocoding-service, because it's not permitted to store geocoded data permanent.
There are free services without this restriction, some keywords to search for: mapquest, nominatim, geonames(but these services are less accurate than google)
Another option would be to use a FusionTable. The geocoding would run automatically(but the daily limits are the same as for the geocoding-service). The benefit: the geocoding is permanent(you can't access the locations directly by e.g. downloading the DB-dump), but you may use the coordinates for plotting markers(via a FusionTablesLayer) or filtering(e.g. by distance)
The number of entries shouldn't be an issue, 100k is no problem for a database
Hello everybody and thank you for taking the time to check out my question. I am a 9th grade teacher making a customized grade sheet and just finished a GUI that lets parents see their kid's grades. However, on the actual gradesheet I use cell notes (previously, comments) to store information specific to a grade (i.e. Jimmy was missing part b of the assignment). I would like to be able to access the data within those notes.
I would like to consider myself a good researcher but have not been able to find a single reference to apps scripts interacting with cell comments or notes. Any help on this would be greatly appreciated.
Also, should there be no infrastructure to support this, I am very open to other ideas on how to store related information while still keeping my interface more more or less constant.
my interface: https://docs.google.com/spreadsheet/ccc?key=0AuptsBCDJBomdGtjc0xBcDctZ3FENmlsdlpRQmdkNkE
parent interface is a simple form that takes a student id number and returns a flex table with grade headers and only that students information.
You can use the Range.getComments() method to gain access to the comments on cells.
https://developers.google.com/apps-script/class_range#getComment
Update (original author: Simon):
Google Docs Comments were relaunched on Sept 5th 2012 as discussion-style comments. All the old Comments were renamed to be Notes. See this notice. The curious thing is that Google has kept the old simple Notes, but not provided any documentation for it.
I am a student at a University. With the placement process going on, we have an internal placement website that shows updates and status about various companies I have applied to. Since the number of companies is too large it becomes cumbersome to scroll through the complete list to find information. Sometimes, I just miss some things. Now, to tackle this problem, here is what I want to do:
The data is in an HTML table. Each row shows information about one company: Some dates, Status(Not/Shortlisted/Applied), Some yes/no options etc. each in a different column. Once I open the page I want to be able to extract information about which companies I got shortlisted in, and in which ones I didn't make it.
What is the right technology to do this ? I am thinking of writing a Greasemonkey user script (I have never actually written any, but how hard could it be ?). What other options do I have?
Edit: I don't quite understand why this question has voted to be closed?
I just displayed a use case for something general: On opening a web page, automatically extracting information from the page and display it to the user. What is the easiest and sufficiently powerful way to achieve this?
Since you can't get access to the website's database, Greasemonkey would be your best automation approach. However, this task is likely to be over before you can get a decent script up from scratch.
Your best practical approach is to save the pages and/or copy and summarize the data in MS Excel, or equivalent.
~~~~~~~~~
Here at SO, We will not develop any but the simplest Greasemonkey scripts for you from scratch (unless they are fun somehow ;) ). But, you can sometimes get such help in the "Script requests forum" at userscripts.org.
In order for someone to help you, they will need:
A clear idea of exactly what data gets manipulated, and how.
Access to the target site. Or access to saved snapshots of the target pages. GM scripts are extremely dependent on the details of the target page.
"other option":
ctrl + F
enter shortlisted
enter
ctrl + G <--repeat last search