Proposal for additional system-wide gestures (swipes and taps)
Introduction
The following proposal aims at making the gesture based user experience of Sailfish OS more consistent throughout the entire system and eradicate the need for “makeshift” solutions in certain apps, that do not fit into the basic concepts of this user experience or hamper the “all screen real estate is yours” paradigm.
Although some of the described gestures might already be existing as patches, having them natively (out-of-the-box) from Jolla in Sailfish OS would be the best.
Swipes FROM the screen edge
In Sailfish OS, we currently have “Swipe from the screen edge towards the screen center” gestures:
- A swipe from the right screen edge takes you to “Home”, with the covers of all your currently running apps.
- A swipe from the left screen edge takes you to “Events”, where you can see your current notifications.
- A swipe from the bottom screen edge takes you to the app grid, from where you can launch apps.
- A swipe from the top screen edge closes the currently used app (when done from either corner), or takes you to the top menu (when done from the middle), where you can switch ambiences, trigger quick actions and toggle settings.
Swipes AT the screen edge
In addition to those “Swipe FROM the screen edge” gestures, this proposal introduces “Swipe AT the screen edge” gestures. They are performed by simply placing one finger at the desired screen edge and moving it along this edge (instead of moving it towards the screen center). Here is what they'll do:
- Right screen edge: One out of two possible actions, depending on whether you currently are in an item view (e.g.: picture, web page, document, etc.) or in a list view (e.g.: gallery grid, message list, artist/album list, etc.):
- Zoom in/out (for item views):
An additional, but much more practical way compared to the well known 2-finger-pinch gesture, because you need only one hand to conveniently do it. Be it that you only have one hand free because you are carrying/holding something with your other hand, or that you cannot use your other hand at all because it is injured/disabled/missing, this gesture greatly improves accessibility. - Aimed scrolling (for list views):
Quickly and precisely scroll to a desired location within the current list with the help of displayed clues. This is especially helpful for very long lists and (partly) eradicates the need for a search or filter feature. When you start moving your finger, the clues will appear (relative to your current location in the list). Continue moving your finger until you are at the desired location (clue), then lift your finger to immediately scroll to that location. The displayed clues depend on the list content. Following are a few examples:
.) Gallery grid: Date on which the picture/video was taken, user-configurable as absolute date (January 2018, August 2018, etc.) or as relative date (6 months ago, 1 year ago, etc.).
.) Message list: Date on which a message has been sent/received. Clues are similar to those of the gallery grid.
.) Artist/Album list, Contact list: Letters of the alphabet.
- Left screen edge: Open top/bottom pulley menu
No need to scroll up/down the entire page in order to access a pulley menu (if one is available). With this gesture, it's immediately accessible no matter where you currently are on the page.
- Bottom screen edge: Actions specific to the app currently being in the foreground
Every app can have different actions here. Following are 5 examples:
- Browser: Go backwards/forwards one page (no need to bring up the navigation bar)
- Gallery: Go to previous/next picture or video (no need to entirely zoom out or scroll to the picture edge)
- Email: Go to previous/next e-mail
- Media: Skip to previous/next track
- Documents: Go to previous/next document page (no need for scrolling to get there)
- Top screen edge: Actions specific to a “chosen” (background) app
The best example for this gesture is the Media app. If you are listening to music while using a different app in the foreground, you can instantly skip to the previous/next track in your playlist without having to switch to the Media app first. The app that receives these swipe events can be chosen in different ways:
- By selecting an app in “Settings” that will automatically get this status once it is started (ideally with the additional option to only do the assignment if no other running app is the currently chosen app).
- By adding a corresponding button to every app cover in the “Close/Rearrange apps” view (the one you get by long-pressing in “Home”), that allows both selecting/switching and deselecting an app.
Double-taps AT the screen edge
By introducing “TAP at the screen edge” gestures, another set of actions is implemented. For the sake of user experience simplicity, this proposal only considers double-taps. Like with the swipes, every screen edge triggers a different action when double-tapped:
- Right screen edge: One out of three possible actions:
- Reset zoom level to default (for item views):
Pretty much straightforward and fits perfectly with the swipe gesture of this screen edge (see above). - Manually enter desired scrolling location (for list views):
Display a date selector, letter selector, etc. (depending on the list content) for even more precise control compared to the swipe gesture of this screen edge (see above). - Alternative function (user-configurable – instead of the two actions described above):
Trigger one of the left screen edge actions (see below).
- Left screen edge: 2 possible actions (user configurable):
- Switch between the 2 last used apps (convenient if you want to do copy-and-paste) or between the currently used app and the “chosen” app (see above).
- Go to the “Permissions” screen of the currently used app (yet to be introduced in Sailfish OS), where you can modify all access permisions for this app on-the-fly as well as check when, how often and how long this app has accessed or attempted to access certain resources of your device, such as your address book or any positioning data (convenient for privacy management and for finding out, if an app deserves your trust).
- Bottom screen edge: Actions specific to the app currently being in the foreground
Again 5 examples:
- Browser: Go to the tab selection screen
- Gallery: Close the current picture/video and go back to the gallery grid
- Email: Close the current e-mail and go back to the containing folder
- Media: Pause/resume playback
- Documents: Close the current document and go back to the document list
- Top screen edge: Actions specific to the “chosen” app
To stay with the example from above, by using this gesture, you can pause/resume playback of the Media app without having to leave your currently used app at all.
Tooltips for screen edge swipes and taps
To assist the user in knowing/remembering which gesture does what, tooltips are used. If the user places one finger at any of the screen edges without swiping or tapping, icons will appear that tell what will happen if the user swipes into a certain direction from the current position or (double-)taps at this position. If the user lifts the finger without having moved it, the icons will automatically disappear and no action will be triggered.
In case of the app-specific actions (bottom and top screen edges), these tooltips can also include additional information, such as for example the title, elapsed time and remaining time of the track currently played by the Media app.
Swipes and taps on "Home"
Last, but not least, this proposal also introduces additional gestures for the app covers.
Currently, we have the following possibilities on “Home”:
- Tapping an app cover switches to that app.
- Tapping the cover button of an app triggers the respective cover button action.
- Long-pressing anywhere on the screen switches to “Close/Rearrange apps” view.
- Swiping left or right from anywhere inside the screen takes you to “Events”.
- Swiping up from anywhere inside the screen takes you to the app grid.
With the exception of “swipe up”, all of these gestures remain unmodified. For the “swipe up” gesture, it now makes a difference if the swipe is done on a free spot inside the screen or on an app cover:
- Swiping up on a free spot inside the screen takes you to the app grid.
- Swiping up on an app cover will display up to 4 additional cover buttons (depending on the app).
A similar differentiation is made for the “swipe down” gesture:
- Swiping down on an app cover dismisses the additional cover buttons (if they were displayed) or closes the app (if the additional cover buttons were not displayed).
- Swiping down on a free spot inside the screen takes you to the top menu.
I like the way it is at the moment. I just found a patch to restore the MeeGo style close of apps. I don'T mind adding more gestures, but I just think it should be optional to the user to decide which let say type or style of gestures should be used.
Nature is pragmatic and as such it is economic - so keep it simple is always a winning strategy and should be also applied here. I already commented on another place about this. At least let me decide at the end what I should use - for example old style vs. new style. Having choice == Having freedom!
thanks
deloptes ( 2019-01-14 02:13:33 +0200 )editWhile the idea is sound, a lot of gestures can confuse users and will require more mental power when using the phone.
addydon ( 2019-01-14 07:20:49 +0200 )edit@addyon not really.
Complex gestures and shortcuts are difficult only for a beginner and will help power users a lot.
juiceme ( 2019-01-14 08:44:35 +0200 )editI think overloading the UI with multiple gestures, making the outcome dependent on where exactly the gesture was performed, is a certain road to failure. If you remember the very first reviews by tech journalists when Jolla 1 came out, many of them just couldn't use the phone because they failed to grasp the concept of "swipe from outside the screen" versus "swipe from inside the screen". The community was more intelligent, but just recently I read here on TJC that somebody claimed he never managed to use the gesture for pulley menus.
In Sailfish 3, the moved into the direction you propose here with "1 gesture, 2 possible results" with both top menu and app close being triggered by the same gesture. This was not a good decision and makes me see a top menu whenever I want to close an app. I'm certain that when moving my finger down the screen, I don't want to pay attention to where exactly it is and if the gesture could be misread as "open pulley menu" when my intention was to scroll through a list.
ossi1967 ( 2019-01-14 11:13:29 +0200 )editI see it a failure if system usage is geared to a direction of the least common denominator; Should we really adopt the philosophy that the dumbest people should dictate how an UI is to be used???
At least there should be "advanced configuration options" that could be activated when an user wants to enable options that could confuse casual tech journalists and other complete idiots.
juiceme ( 2019-01-14 11:47:31 +0200 )edit