System.EdgeGesture.DisableTouchWhenFullscreen can be used in a desktop app to disable edge swipes, but has no effect on a Store App. We will be using a Store App on a kiosk - is there any way to stop the edge swipes, charms bar app closing gestures?
Related
I'm trying to run the attached project on Windows 10 Pro (latest version available without Windows Insider Program).
Basically it is a fullscreen browser window that navigate to http://www.google.com.
I configured Windows in Tablet mode, in order to let the touch keyboard popup whenever any text field in the page (in this case the query one) gets focused.
Then, I packaged the application with electron-windows-store in order to let electron work as Windows Store application.
When I start the application and Google home page is loaded I'm not able to use the touch keyboard, because it pops up but immediately disappears, like if electron tries to acquire again focus and causes touch keyboard disposal.
I tried also to disable fullscreen mode and setup frame coordinates in order to be as it was in fullscreen, but no success.
Any suggestion?
TestApp.zip
GitHub Repo
This seems to be related to an open issue on the Electron GitHub repository. You might have to wait for the Electron team to introduce this improvement.
I've managed it. The issue was caused by an old dependency to electron. Once updated it to the latest version I know (1.4.7) it all started working.
Trying to employ kiosk based on Chrome browser. Just Installed Kiosk app extension, everything is fine, but in Chrome's application mode there are no toolbar. I basically need "Back", "Forward", "Home" buttons. Is there any way to enable toolbar in app mode, or add it when extension creates window?
Navigation is not meaningful for Chrome Apps. It is, in fact, disabled.
So if your app changes state, you need custom controls for that state anyway.
If you have embedded web content in the app in a <webview>, then you need to make your own custom controls for that. See the browser app sample.
I'm trying to develop a kiosk web application that uses Google Chrome on kiosk mode setup which loads automatically after start-up.
http://www.sitepoint.com/google-chrome-kiosk-mode/
The kiosk web application also uses a virtual keyboard plugin for Google Chrome for the text inputs.
http://xontab.com/Apps/VirtualKeyboard
I'm planning to setup a computer unit with a touch screen monitor for the kiosk.
Note: It's my first time to develop a web application that uses the kiosk mode setup for Google Chrome and I don't have a touch screen monitor for testing. I wanted to ask this question for developers that has experience with this.
My question is:
Does Google Chrome on kiosk mode setup automatically detects my touch screen monitor?
Does Google Chrome automatically enables touch features when my web application is on kiosk mode such as swipe for scrolling up and down.
A touch screen is an input device just like a computer mouse - Google Chrome receives touch events the same way it receives mouse events (although the events are different).
The annoying thing when starting working with touch screens is that the standard click event that we are used to, is triggered after a delay comparing to mouse event. You should listen for tap event or use a library such as https://github.com/ftlabs/fastclick
Making long story short developing a kiosk application integrated with a touch screen is similar to developing mobile websites. You should probably use a JavaScript library to support all kind of touch events such as tap, swipe etc. See http://hammerjs.github.io/
You may also find this website useful http://peter.sh/experiments/chromium-command-line-switches/
I know this answer is six years later but for anyone reading this. You can run Google Chrome in Kiosk Mode with custom options. This can be done by creating a new account and right clicking (in windows 10) on the icon for this account.
Under properties add the following tags to optimize the application for touchscreen use:
–-touch-events –enable-viewport
It appears Firefox OS has a way to turn a device's camera flash on and off.
Is there a way to turn the camera flash of a phone on or off in a standards-compliant way from within web browsers running on the phone? E.g. within Chrome on Android/iOS, Safari on iOS, etc.?
It appears you can read the capabilities but not manually set them.
Asking to see if it's possible to build a mobile webapp that works like a flashlight.
I just started developing wp8 apps coming from android. What is the equivalent UI control for an app drawer (android) to be able to quickly change from page to page? I'm looking for the control used to change from page to page in OneNote
EDIT: I found out that Windows phone 8.1 offers flyouts which can be made to look like an app drawer.
The equivalent to the android app drawer is probably the windows phone appbar, but the appbar in windows phone is not used for navigation - only settings like actions.
In windows phone, navigation is done primarily through your own interface, with heavy use of the BACK button to go to previous pages. The main class you use in windows phone is the NavigationService
I don't believe the OneNote uses standard UI components, so you would probably have to create your own
There are plenty of solid resources online that should help. Start here:
http://msdn.microsoft.com/en-us/library/windowsphone/develop/ff626521(v=vs.105).aspx