Understanding Players’ Interaction Patterns with Mobile Game App UI via Visualizations
Abstract.
Understanding how players interact with the mobile game app on smartphone devices is important for game experts to develop and refine their app products. Conventionally, the game experts achieve their purposes through intensive user studies with target players or iterative UI design processes, which can not capture interaction patterns of large-scale individual players. Visualizing the recorded logs of users’ UI operations is a promising way for quantitatively understanding the interaction patterns. However, few visualization tools have been developed for mobile game app interaction, which is challenging with multi-touch dynamic operations and complex UI. In this work, we fill the gap by presenting a visualization approach that aims to understand players’ interaction patterns in a multi-touch gaming app with more complex interactions supported by joysticks and a series of skill buttons. Particularly, we identify players’ dynamic gesture patterns, inspect the similarities and differences of gesture behaviors, and explore the potential gaps between the current mobile game app UI design and the real-world practice of players. Three case studies indicate that our approach is promising and can be potentially complementary to theoretical UI designs for further research.
1. Introduction
Understanding players’ interaction patterns in smartphone devices with mobile game apps is important for game experts to develop and refine their app products (Punchoojit and Hongwarittorrn, 2017; Folmer, 2007; Bunian et al., 2017). For example, inspecting players’ gesture sequences in a gaming app can help its game user experience (UX) researcher determine the players’ gaming expertise and understand the players’ engagement; understanding how players interact with mobile multi-touch devices can help the user interface (UI) designers verify if the intended UI of the app is convenient enough for the gamers, thus enabling appropriate modification and optimization of the original UI design.
Conventionally, game experts study players’ interaction patterns with mobile game apps through intensive user studies with target players or iterative UI design processes to achieve their purposes (Korhonen and Koivisto, 2006; Park et al., 2013; Hellstén, 2019). For example, UX experts would pay attention to the players’ performance by extracting and analyzing their in-game behaviors to understand whether they are fully engaged in the game; UI design experts mainly propose mobile game UI designs based on some theoretical experiments and comparison with the designs of competitor products, trying to provide more user-friendly interface designs and ensure a smooth players’ interaction. While these evaluation methods are informative and can potentially help identify the general patterns of the players’ gaming interactions via mobile app UI, they could not provide quantitative understanding of user interactions with the app product itself.
One alternative way to assist game experts to study players’ interactions with mobile game apps is to directly record the user interaction data from mobile devices by developing logging and analysis tools. Visualization is a promising way to enable game experts to dig into deeper the interaction data (Harty et al., 2021; Froehlich et al., 2007; Krieter and Breiter, 2018). However, most existing visualization tools have long been developed for desktop-based interaction data and cannot be simply borrowed and applied in mobile game scenarios (Li et al., 2016, 2019). It becomes more challenging when the mobile game app supports intuitive direct-manipulation and multi-touch operations, e.g., the joysticks moving towards multiple directions, i.e., up, down, left, right, and different skill buttons possessed by the game character in the app, mimicking real-world metaphors. There have been some works seeking to measure and analyze low-level interaction metrics on mobile touch screen devices by inviting participants to perform various controlled tasks (Anthony et al., 2012, 2015). Nevertheless, most of them focus on aggregating primitive metrics such as task completion time and accuracy/error rate, which do not always capture high-level interaction issues that players may encounter in real-world usage scenarios. Some studies present visualization methods for low-level interaction logs to identify noteworthy patterns and potential breakdowns in interaction behaviors, e.g., elderly users’ interaction behaviors (Harada et al., 2013), however, they target at a few specific application interfaces such as Phone, Address Book and Map that do not require dynamic user operation. For mobile game that supports multi-touch options (e.g., with joysticks), players usually have more dynamic and complex interactions with the app UI, making it more challenging to visualize the interaction and identify the patterns behind.
To fill this gap, in this work, we present a visualization approach that aims to help game experts understand players’ interaction patterns in the context of a multi-touch gaming app with joysticks and skill buttons and further improve its UI design. Particularly, we first identify players’ dynamic gesture patterns that correspond to a series of interaction data. Then, we develop a visualization tool to help inspect the similarities and differences of gesture behaviors of players with different gaming expertise and explore the potential gaps between the current mobile game app UI design and real-world practice of players. We evaluate our approach on three cases (i.e., interaction skill comparison, individual interaction skill, and user interface design verification). Both qualitative feedback and quantitative results of case studies suggest that our approach is promising and can be complementary to mobile game players’ behavior understanding and theoretical mobile game UI designs.
Our contributions can be summarized as follows: 1) a gesture-based logging mechanism that comprehensively records user interaction and identifies players’ gesture patterns; 2) a novel visualization approach that identifies the similarities and differences of high-level gesture behaviors on a touchable mobile device, and 3) case studies that provide both quantitative and empirical evidence to verify the efficacy of our approach and elicit promising UI design implications. In the following sections, we briefly survey the background and related work, followed by an observational study to identify the mobile interaction data characteristics and design requirements for the proposed visualization approach. Then, we carry out three case studies to verify our approach. Finally, we conclude our work with discussions and limitations and shadow the potential design implications discovered through our approach.
2. Related Work
Literature that overlaps with this work can be categorized into three groups: mobile interaction data analysis, assessment of mobile UI design, and gesture data analysis.
2.1. Mobile Interaction Data Analysis
User behavior analysis has been intensively studied in the game domain (Li et al., 2016; Kang et al., 2013; Lee et al., 2014). Li et al. analyzed players’ actions and game events to understand reasons behind snowballing and comeback in MOBA games (Li et al., 2016). These studies mainly focus on in-game players’ behavior analysis instead of their interaction with the game devices. The most similar studies come from web search (Blascheck et al., 2014; Kim et al., 2015; Huang et al., 2012), where statistical analyses of e.g., eye-tracking or mouse cursor data mainly provide quantitative results. Visualization techniques are developed to allow researchers to analyze different levels and aspects of data in an explorative manner, which can be categorized into three main classes, namely, point-based, area-of-interest-based, and approaches that combine both techniques (Blascheck et al., 2014). Among these methods, the aggregation of fixations over time or participants is known as a heat map that summarizes illustrations of the analyzed data and can be found in numerous publications. However, many methods proposed for desktop website analysis cannot be simply applied to mobile game apps since these methods are based on single-point interaction such as mouse or eye movements while mobile devices support intuitive direct-manipulation and multi-touch operations (Humayoun et al., 2017).
Timelines are frequently used to represent the interaction information. Previous solutions have provided static and limited representations of them (Carta et al., 2011). A new solution was designed to represent and manipulate timelines, with events represented by a level and a small colored circle (Burzacca and Paternò, 2013). It also includes vertical black lines among events that indicate a page change, which provides effective interactive dynamic representations of user sessions. Nebeling et al. presented W3Touch to collect user performance data for different mobile device characteristics and identify potential design problems in touch interaction (Nebeling et al., 2013). Guo et al. conducted an in-depth study on modeling interactions in a touch-enabled device to improve web search ranking (Guo et al., 2013). They evaluated various touch interactions on a smartphone as implicit relevance feedback and compared these interactions with the corresponding fine-grained interactions in a desktop computer with a mouse and keyboard as primary input devices. Bragdon et al. investigated the effect of situational impairments on touch-screen interaction and probed several design factors of touch-screen gestures under various levels of environmental demand on attention compared with the status-quo approach of soft buttons (Bragdon et al., 2011). To date, however, few empirical studies have been conducted on mining touch interactions with mobile game apps to understand players’ behaviors and further suggest mobile application UI designs via a visual analytics approach.
2.2. Assessment of Mobile UI Design
Many mobile applications, particularly game apps such as ARPG (Action Role Playing Game), introduce gesture operation to control the game character in the game app freely (Hesenius et al., 2014; Ruiz et al., 2011). However, no single gesture operation can resolve all the interaction issues in the mobile game application scenarios due to different screen sizes, individual behaviors, and the form of different controls used. Besides, gesture operation requires extensive learning. Thus, most existing mobile game applications, particularly role-playing games, adopt a virtual joystick and skill buttons for players’ interaction (Baldauf et al., 2015).
Previous studies that focus on the assessment of mobile UI design can be summarized into two categories, qualitative methods and quantitative methods. For example, when designing mobile app interfaces, targets are generally large to make it easy for users to tap (Anthony, 2012). The iPhone Human Interface Guidelines of Apple recommend a minimum target size of pixels wide * pixels tall (Clark, 2010). The Windows Phone UI Design and Interaction Guide of Microsoft suggests a touch target size of pixels with a minimum touch target size of pixels * pixels. The developer guidelines of Nokia suggest that the target size should not be smaller than cm * cm or pixels * pixels. Although these guidelines provide a general measurement for touch targets, they are inconsistent and do not consider the actual varying sizes of human fingers. In fact, the suggested sizes are significantly smaller than the average finger, which may lead to many interaction issues for mobile app users. Hrvoje et al. presented a set of five techniques, called dual finger selections, which leveraged recent developments in multitouch sensitive displays to help users select extremely small targets (Benko et al., 2006). The UED team of Taobao researched to determine hotspots and dead-ends and to identify the control size using the thumb111http://www.woshipm.com/pd/1609.html; they concluded that the minimum target size should be mm by single thumb operation to achieve an average accuracy higher than 95%. However, most existing works have focused on the vertical screen mode by single-hand operation, and only a few have discussed the landscape mode using both hands. In this work, we focus on the assessment of virtual joystick and skill buttons by leveraging players’ interaction data with the mobile game app to analyze the triggering and moving ranges of virtual joystick and skill set areas.
Regarding the quantitative methods, researchers have been working on modeling user perception and subjective feedback of user interface, e.g., the judgment of aesthetics (Miniukovich and De Angeli, 2015), visual diversity (Banovic et al., 2019), brand perception (Wu et al., 2019), and user engagement (Wu et al., 2020). Typically, a set of visual descriptors would be compiled to depict a UI page in terms of e.g., color, texture, and organization. Specifically, user perception data are collected at scale and corresponding models are constructed based on some hand-crafted features (Wu et al., 2019). However, feature engineering cannot ensure a comprehensive description of all the aspects of UI. Recently, deep learning has demonstrated its decent performance on learning representative features (Krizhevsky et al., 2012). For example, a convolutional neural network is adopted to predict the look and feel of certain graphic designs. Wu et al. leveraged deep learning models to predict user engagement level on animation of user interfaces (Wu et al., 2020). Similarly, perceived tappability of interfaces (Swearngin and Li, 2019) and user performance of menu selection (Li et al., 2018a) can also be predicted with the assistance of deep learning methods. However, the existing studies provide prediction scores of user perception towards different UI designs while in our work, we study how players experience in real-world practice via the current design of mobile game UI by visually analyzing their interaction patterns and shadow the design implication of mobile game UI interface.
2.3. Gesture Data Analysis
Owing to the pervasiveness of multi-touch devices and wide usage of pen manipulation or finger gestures, a great number of researchers have conducted on stroke gestures analysis generated by users (Zhai et al., 2012; Tu et al., 2015; Vatavu and Ungurean, 2019; Leiva et al., 2018). Most of the related existing works focus on gesture recognition, which matches gestures with target gestures in the template library based on their similarity. For example, Wobbrock et al. developed an economical gesture recognizer called $1, which is easy to incorporate gestures into UI prototypes (Wobbrock et al., 2007). They employed a novel gesture recognition procedure, including re-sampling, rotating, scaling and matching without using libraries, toolkits or training. Yang Li developed a single-stroke and template-based gesture recognizer, which calculates a minimum angular distance to measure similarity between gestures (Li, 2010). Ouyang et al. presented a gesture short-cut method called Gesture Marks, which enables users to use gestures to access websites and applications without having to define them first (Ouyang and Li, 2012). In order to understand the articulation patterns of user gestures, Anthony et al. studied the consistency of gestures both between-users and within-users and some interesting patterns have been revealed (Anthony et al., 2013). Some works conduct research on different users, such as children (Anthony et al., 2012, 2015), elderly people (Lin et al., 2018), and people with motor impairments (Vatavu and Ungurean, 2019), aiming to improve user experiences on mobile device interactions.
In this work, instead of gesture recognition, we focus on analyzing user behaviors to find similar and preferred stroke gestures, i.e. similar operations generated by players when playing games. Considering different features of stroke gestures in terms of position, direction, shape and so on, we cluster stroke gestures to reveal users’ common behaviors by resampling a stroke gesture as a point path vector, followed by a definition of a distance function and clustering algorithms.
3. Observational Study
To understand the mobile game app interaction data characteristics and identify the design requirements of our visualization approach, we worked with a team of experts from an Internet game company, including two UX analysts (E.1, male, age: and E.2, female, age: ), one data analyst (E.3, male, age: ) and two game UI designers (E.4, female, age: and E.5, female, age: ), a typical setup for a game UX team in the company. All of them have been in the game industry for more than two years. To obtain an understanding of players’ interaction patterns and experiences with the mobile game app, the game experts have different responsibilities. Particularly, E.1 and E.2 would conduct two main approaches, namely, in-game interaction data analysis with the help of E.3 and subjective interview with the testing players to understand their interaction patterns and provide UI design suggestions for E.4 and E.5. To obtain detailed information of the current practice of the game experts, with consent, we shadowed the team’s daily working process, including videotaping how they observed players experiencing the game, conducted testing experiments, and on-site interviews with the players. Later, we carried out retrospective analysis together with the game team on their conventional practices.
Participants. The game experts recruited participants ( female, avg. age: ) from a local university. They were undergraduate and postgraduate students. Some of them are novice players, ensuring that the study of players’ interaction applies to different expertise of target users.
Procedure. In the prior study, the participants were first asked to complete a task by a mobile app similar to the game of Whack-A-Mole222https://wordwall.net/about/template/whack-a-mole using a mobile phone device with the size of cm * cm and the resolution of * p. Particularly, the mobile screen is divided into multiple small squares, of which a random square turns red, requiring the participants to accurately click on within second. In this experiment, the size of the square and its position were randomly combined, and the participants needed to click times consecutively and this lasts for about one and a half minutes. The participants held their hands horizontally and clicked on the screen with their thumbs. The square size was designed as a variable with the side length taking from mm to mm. The keystroke time is defined as the time from the appearance of the red square to the time when the player clicks on the screen and the game experts calculated the distance from the red square to the lower-left corner of the screen. The game experts distinguished the square size and position for data statistics.

Result Analysis. To match players’ real-world usage scenarios and regarding the size and pixel adaption of different mobile phones, the game experts take “mm” as the measurement unit instead of pixel. As shown in Figure 1(c), the most comfortable zone for the left thumb’s clicking (i.e., the shortest reaction time) is the area with a radius of – mm. Considering the occlusion of the combat interface when the joystick is moved to the middle of the screen, the UI designers maintained that the secondary comfort zone ( – mm) is recommended as the design area. The designers further converted to the distances between the center of the joystick and the left and bottom of the screen, i.e., mm – mm. The experts also investigated other game competitors’ UI designs and identified that most of the center of the joystick keeps at least a mm distance from the left and lower corners of the screen (Figure 1(d)). Note that the results are based on the randomly appearing square button and the focus area is the recommended design area, i.e., the lower left part of the screen when holding the phone horizontally. Specifically, when the side length is set to mm, the accuracy of square clicking333click accuracy: in a single response, participants correctly hit the target as the end. If the number of hit is greater than 1, the response is considered to be a failure. That is, the click accuracy is the percentage of the number of times that the finger successfully hit the stimulating red square for the first time to the total number of the red squares can be larger than % (Table 1). The joystick size of game competitors is basically consistent with experimental results, i.e., between and mm.

Similarly, the UX experts found that the most comfortable zone for the right thumb’s clicking (i.e., the shortest reaction time) is also the area with a radius of – mm, as shown in Figure 2(a). And the secondary comfort zone, i.e., – mm is also recommended as the design area. Although – mm is also a comfortable zone for players to operate, it can easily obscure the screen. UI design experts suggested that as the mm arc is close to the edge of the screen, some heavy skills can be placed here, e.g., ultra-low frequency skill buttons. The game experts also surveyed other game competitors’ low-frequency skill buttons and concluded that their positions are close to the mm arc, i.e., the most comfortable area (Figure 2(b)).
Following a similar procedure, the game experts found that when the thumb’s clicking range is within the comfort area recommended above, 90% accuracy can be ensured if the diameter of the Normal Attack button is over mm and the diameter of the other skill buttons is above mm (Table 1). The survey of game competitors also confirms that the diameter of skill buttons is within to mm and the Normal Attack button is within to mm.
Size (mm) | 7 | 8 | 9 | 10 | 11 | 12 | ||
---|---|---|---|---|---|---|---|---|
|
77.3% | 77.8% | 94.4% | 85.7% | 87.5% | 100.0% | ||
|
63.0% | 90.6% | 92.3% | 90.5% | 80.0% | 100.0% |
The results of the above experiments provide some initial guidelines for the UI design of their mobile game app. However, UI design experts (E.4 – 5) commented that “although in general, they are consistent with the survey results of other game competitors, they can be quite rough and general.” E.1 further commented that “this experiment requires a high degree of concentration but in reality, players are playing the game in a more relaxed mood.” In other words, the interaction characteristics of a certain mobile game app are not fully considered during the testing when the participants were conducting their actions and they are in a state of tension. UI design experts, therefore, envisioned a more customized and natural way to learn the players’ interaction patterns with the mobile devices for inferring the design suggestions for the joystick and skill buttons. On the other hand, UX experts mainly studied the players’ in-game performance. However, in-game metrics reflects the level of players’ performance but cannot explain the reasons behind their performances, “while a bad interaction with the game app would certainly lead to poor performance of players, this observation cannot be easily captured by the in-game metrics,” said E.3. That is, a good way to identify similarities and differences among players’ interactions with the mobile game app is still missing for the UX experts.
To ensure that the ontological structure of our approach fits well into domain tasks, we interviewed the game experts (E.1 – 5) to identify the experts’ primary concerns about the analysis of players’ interaction patterns with playing the mobile game app and potential obstacles in their path to efficiently obtaining insights. At the end of the interviews, the need for a gesture-based visualization approach to ground the team’s conversation with assessing mobile interaction patterns emerged as key themes among the feedback collected. Despite differences in individual expectations for such approaches, certain requirements were expressed across the board.
R.1 Distinguishing gesture behaviors from a spatiotemporal perspective. Conventionally, the game experts leveraged heat map visualizations to observe the distribution of clicking spots, regardless of interactions caused by different fingers’ interactions, which cannot show the quantitative information. Furthermore, the collective heat map distribution cannot provide more details of different gesture behaviors that may occur at different timestamps and positions, failing to shed lights upon players’ interaction patterns. Therefore, the game experts wished to distinguish different gesture behaviors from a spatiotemporal perspective.
R.2 Inspecting behavior differences and similarities among players. One concern of the game experts was that the interaction patterns cannot be easily inspected only by leveraging the in-game metrics. For example, the game experts wished to know “what the common interactions with the mobile device is among the players and what the difference is”, thus allowing them to understand how the operation skill may influence their in-game performance, which can be complementary to the performance of in-game behaviors.
R.3 Identifying interaction areas with appropriate scales and positions. UI design experts typically focus on three aspects of designs, i.e., style design, scale design, and position design. While style design is reflected in the interactive elements, caters to the gameplay experience, and has sufficient feedback to the player’s operation behavior, players’ interaction patterns can be largely influenced by the scale and position designs. Therefore, the game experts, especially the UI design experts, wanted to identify the appropriate interaction areas in terms of scales and positions that can reflect the real-world gameplay interaction experiences accurately.

4. Approach
In this section, we first illustrate how we collect the gesture-based interaction data and then introduce our visualizations to help game experts understand the mobile game app interaction patterns.
4.1. Gesture-based Logging Mechanism
We have developed an application-independent Android program that can interact with the mobile OS. By detecting every touch event and retrieving the screen coordinates with the timestamp of the touchable screen, we can log all the touchable-screen actions through multiple functions, e.g., ACTION_DOWN (touch the screen), ACTION_MOVE (move on the screen), and ACTION_UP (leave the screen). We consider the players’ interaction data as high-level gestures, which are the trajectories of players’ fingers on the multi-touch screen that can be recorded as a series of points generated by the same finger of the player in one session of interaction. Particularly, when the player places a finger on the screen, a starting point is recorded. The player then moves this finger to other places, generating several corresponding points (, , …, ). When this interaction session terminates, the player raises his/her finger and ends the recording of the gesture. We define the gesture = [, , …, , …, ], where point is described in terms of (, ) with corresponding timestamp indicating the corresponding time-lapse from the beginning. Each gesture trajectory has an associated id, which records the current session the program identifies during the interactions.
4.2. Gesture-based Visualization
The basic design principle behind our approach is leveraging or augmenting familiar visual metaphors to enable game experts to focus on analysis (Li et al., 2018b). Considering the gesture data being analyzed and the above-mentioned requirements, we visually encode the data in a manner that ensures that the patterns and outliers are easily distinctive without overwhelming the analysts.
We define a gesture as a series of finger actions, typically starting with a “finger down”, followed by several “finger moving”, and ending up with “finger lifting”. We adopt a timeline-like metaphor to align all the actions of the fingers in a radial layout, presenting the overview of the distribution of finger actions intuitively, as shown in Figure 3. We choose a radial layout because it allows users to analyze high-level interactions (e.g., periodic behaviors) and compare the gesture patterns in different stages in a more concentrated manner. Meanwhile, the mobile game interface could be placed inside the radial circles, helping analysts better link temporal events with the corresponding interaction dots and spatial trajectories on the screen. We apply three concentric circles to denote the three fundamental actions of fingers, namely, “finger touching” in the most inner circle, “finger moving” in the second inner circle, and “finger lifting” in the outermost circle (R.1). The spatial position of each finger action is determined by its timestamp and the entire recorded period is considered in a clockwise direction. One complete and continuous gesture is represented by a Bezier curve that links the starting action (finger touching) and the ending action (finger lifting) with a series of finger movements distributed between the two actions. We also encode gesture durations into the parameters of two controlling points of the Bezier curve, i.e., the height of the curve corresponds to the related gesture duration. We also provide heat map visualization as a qualitative complementary to the proposed gesture visualization.
In addition, to encode the gestures to concentric circles, we filter out gesture trajectories based on spatial locations, e.g., we can visualize the trajectories that correspond to a particular area of the mobile app interface (R.1). Particularly, we design a spatial-based query to link the trajectories to the corresponding finger actions along the concentric circles (see the selected area). We directly integrate the UI design into the visualization. On the left side of the view, the interaction within the area controls the orientation and motion of the game character in the app, and the right side, i.e., the five rounded areas correspond to different skills of the game character. Figure 3(b) presents the use of the spatial-based query for a relatively long duration of operation, which corresponds to a gesture trajectory. Most of the finger movements occur on the left side, whereas quick taps happen on the right side.
Since the skill button interaction represents certain semantic information, i.e., different skill button indicates different operations on the game character, we can define the high-level interactions in the skill semantic space. As shown in Figure 3(c), analysts can self-define a region, e.g., a circle to surround an area and assign certain semantic information to the area, and the corresponding finger actions within the area would be automatically mapped onto the semantic axes on the outermost semantic circles, which supports the interactive exploration from the skill semantic perspective (R.2).
Conventionally, to identify the typical interaction patterns on the mobile game app screen, the heat map is intensively used. One obvious advantage is the lack of quantitative feedback by using a heat map that only conveys a sense of qualitative density information. After discussion with the game experts, we propose an interactive clustering method based on the interaction data (R.3). Particularly, we allow analysts to select the original interaction dots with a certain radius and the system automatically calculates the clustering center of the selected interaction dots and choose the area with different confidence coefficients, i.e., identify the most appropriate center and size of the area that meets certain confidence coefficients. In this way, a new clustering center could be generated. We further sort the interaction dots based on the distance between and the interaction dot and the longest distance is considered as the new radius (Figure 4).
h
Once we determine the region for a certain confidence coefficient, we then extract the gestures within the region to understand the underlying semantics of the gestures and reveal the common interaction behaviors of players. Note that different gestures contain various number of interaction dots, we need to resample gestures to ensure that all the gesture trajectories are directly comparable. Given a gesture , we assume its original number of interaction dots is and the objective is to sample interaction dots from with evenly spaced sampling. Specifically, the original gesture is represented as and and after sampling, we get a new gesture , where and . They are subjective to
(1) |
In other words, we add the first interaction dot , and then add a new dot sequentially until covers dots, followed by adding the last interaction dot. Those new interaction dots is generated from the original gesture by using linear interpolation. The next step is to measure the distance between two gestures. Particularly, given two gesture sampling vectors: , , and considering the absolute distance and directions, we combine two distance functions, i.e., Euclid Distance and Cosine Distance with adjustable weights, and use K-Means to all sampling gesture vectors. The distance functions are defined as Euclid Distance:
(2) |
and Cosine Distance
(3) |
where
(4) |
and
(5) |
5. Case Study
To evaluate our approach, we conduct several case studies, in which the previous participants are asked to play a mobile game to collect their interaction data for our analysis. Then, we present our visualization approach to the game experts to evaluate the efficacy of our design. Particularly, the evaluation is conducted in the following three cases to examine whether our visualization approach fulfills the aforementioned requirements.
5.1. Participants and Procedures
The education backgrounds of the above-mentioned participants range from computer science, electronic engineering to art designs. We collect information about the participants’ mobile phone usage, including their mobile phone operating systems, the size of their mobile screens, and the frequency of playing mobile games per week. All of them have the experience of using both IOS and Android mobile operating systems. The size of the mobile phone screens they use ranges from to inch to above inch. Regarding mobile game experiences, most of the participants (80%) play games for fun, usually – days per week. Three of the participants consider themselves experts in playing mobile games, five have intermediate-level gaming experience, and the others are novices. Their gaming expertise is based on the number of mobile game apps they have ever played similar to our testing mobile game app and all of them have no prior knowledge of our mobile game app, i.e., nor have they seen it or played with it.
The mobile game app we choose for our study is a type of ARPG, where the player controls the actions of the main game character immersed in a well-defined virtual world and resists attacks from in-game monsters. The orientation and movement of the main game character are controlled by the player through a virtual joystick placed on the left side of the mobile screen, while the skill release controls are placed on the right side of the screen, usually consisting of five skills (i.e., one Normal Attack in the right-button corner surrounded by the other four skills). We have studied different mobile applications (e.g., “address book”, “2048”, “Angry Birds”, “Temple Run”) and identified that they all share the same set of basic down/move/up actions. However, “Angry Birds” only involves limited events of finger move (e.g., launch a bird) and finger down (e.g., make birds explode). Therefore, we choose the virtual joystick mobile game that involves lots of finger down/move/up events fully engaging players via interactions with both hands and has a proper duration ( – min on average) and different levels of difficulty. Thus, the resulting gesture interactions are diverse enough for our experimental analysis.
5.2. Case One: Interaction Skill Comparison
The objective of the first case study is to differentiate novice players from the expert ones based on their interaction data with the mobile multi-touch screen when playing our testing mobile game app. For the first case study, we recruited two novices (one female student, age: and one male student, age: ) and two expert gamers (male students, age: and , respectively) to compare their interaction skills by using our approach. A single gaming session was conducted with each participant for minutes. They were firstly given a brief overview of the basic operating rules of the testing mobile game app. Then, each participant played for three consecutive rounds, and we only collected the operation logs of the last two attempts. The first attempt served as a training session to help the participants familiarize with the testing mobile game app.
h
As shown in Figure 5(1), we applied our visualization approach to the interaction data of a novice player and identified that the player has only three finger movement trajectories, indicated by (a), (b), and (c), respectively. Among the three movement trajectories, (b) and (c) occur within the first seconds, followed by a long-lasting movement of over seconds that continues until the end of the game session. We then observed how the player interacted with the skill buttons by the five corresponding skill concentric semantic circles. Most of the five skill buttons are triggered simultaneously in clockwise order with the Normal Attack (the biggest button) triggered first followed by the other four skills, as indicated by the green rectangles. Since the four skills have several seconds for cooldown, they cannot be triggered immediately if they have been triggered and the Normal Attack is frequently used. We can conclude that the operation of the novice player involves a continuous movement on the left side of the screen to control the orientation and movement of the game character in the mobile game app and consecutive release of the five skills in a clockwise direction. The left (orientation and movement) and the right operation (skills) are not combined well. In other words, the player did not have a good strategy to combine the orientation control and the movement of the game character effectively with the other five skills to defend the game character against attacks.


For comparison, we visualize another expert player’s interaction data, as shown in Figure 5(2). Following the same approach, we first focused on the long-last movement trajectories, which are defined by relatively long and high curves that connect the corresponding “starting” dots (finger down) and “ending” points (finger up). We identified that at the timestamp of each long-lasting movement, the skill set on the right side of the screen is triggered in a regular pattern, i.e., the Normal Attack is typically accompanied by the other four skills that require a certain cooldown time. Thus, the efficient combination of the long-lasting movement to control the orientation of the game character and the release of the skill set results in the player winning the game.
After comparison, the game experts maintained that the four skills, except for the Normal Attack, are lower in operations because of that “a good operation focuses more on joystick control and movement.” (E.1) By contrast, a weak operation depends significantly on the Normal Attack compared with the other skills, and involves fewer finger movement on the left side of the mobile screen. From the visualizations, the game experts commented that “the weak operation continuously utilizes the Normal Attack during the finger moving and touching operations” (see the most inner semantic circle in Figure 5(1)) “while the good operation applies the Normal Attack regularly and intermittently”.
5.3. Case Two: Individual Interaction Skill Improvement
In this case, we demonstrate how the game experts leverage our visualization approach to track the potential improvement of an individual player’s skills when he/she takes a series of attempts of playing a mobile game. The game experts invited another two novices players (a male student with the age of and a female student with the age of ) and asked them to carry out a series of attempts consecutively. Particularly, the participants were first introduced to get familiar with the basic operations and then played the mobile game several times. After finishing the attempts, their interaction data were recorded each time and the game experts conducted an interview on each of them, separately taking a note of each interviewee’s response.
Figure 6 shows a case where a participant (male, age: 28) played the game four times, of which the first two attempts failed to accomplish the game task. To be specific, in the first two attempts, although the participant applied all the skills, they were conducted successively. We observed that continuous touches occur on the second, third, and fourth skill buttons, which should be avoided since that except for the Normal Attack, all the other four skills have a cooldown time that disables the corresponding skill for a certain duration. Another phenomenon that can be witnessed is that nearly all the skill triggering happens simultaneously and there is no correlation between the movement control by the joystick and the skill operations. On the other hand, the last two results correspond to successful attempts. The combination of movement control and skill release becomes more obvious. The triggering time intervals between different skills are elongated, as shown in Figure 6 (d). Furthermore, the movement trajectories are denser and more concentrated in Figure 6 (d) than that in the previous attempts. “In the beginning, I was quite nervous and did not know how to handle so many monsters so I just click on every skill buttons,” said the participant, “I soon realized that without any strategy I would never win the game”, so he began to coordinate his both hands.
Another participant made three attempts. In the first two attempts, the participant also failed to finish the gaming task. Particularly, in the first attempt, the participant only used Normal Attack and the other two skills. Although the movement was quite intensive, there is no combination with the application of skill release. In the second attempt, the participant began to use more skills, but in a random way. Taking a close look at the triggering timestamp of the skills and movement control, the game experts cannot identify any correlation. In the last attempt, the participant commented that “I feel like knowing how to play” when he began to pay attention to the combination of the joystick with the application of different skills. For example, when there was a relatively long-lasting movement, some skills were released immediately. As shown in Figure 7, the ratio of Normal Attack significantly drops, and more operations are given to the joystick and are distributed in the four skills evenly.
Confidence Coefficient: 95% | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
joystick | original radius | sampling number | original center | original number | new center | new radius | new number | distance to the left | distance to the bottom | diameter |
touch | 421 | 8908 | (1462.172, 216.930) | 1228 | (1467.603, 214.247) | 146.331 | 1166 | 15.86 mm | 13.85 mm | 15.61 mm |
move | 444 | 87223 | (1468.190, 237.092) | 86174 | (1475.237, 234.957) | 232.275 | 81865 | 15.37 mm | 15.18 mm | 30.02 mm |
Confidence Coefficient: 99% | ||||||||||
joystick | original radius | sampling number | original center | original number | new center | new radius | new number | distance to the left | distance to the bottom | diameter |
touch | 421 | 8908 | (1462.172, 216.930) | 1228 | (1462.172, 216.930) | 196.650 | 1215 | 16.21 mm | 14.02 mm | 25.42 mm |
move | 444 | 87223 | (1468.190, 237.092) | 86174 | (1470.297, 236.474) | 321.171 | 85312 | 15.68 mm | 15.28 mm | 41.51 mm |
5.4. Case Three: UI Design Verification
This case focuses on identifying the common behaviors of players’ interaction with the mobile game app and further verifying the current UI design of the mobile app interface in real usage scenarios. To approach this, the game experts covered the main operation regions of the mobile game app interface to ensure that the participants operate based on their own habits for an objective evaluation. As shown in Figure 1(b), the size of the covered region is approximately cm * cm. The game experts invited all the participants and asked them to use only two thumbs to operation on the game avatar in the mobile game app for five rounds, no matter success or failure for accomplishing the gaming task. During the gaming process, all the participants were asked to keep their thumbs in a natural curve and use their finger bellies to touch the screen while maintaining the stability of the mobile devices. The mobile device used has a width of cm and a height of cm. The resolution of the device was * pixels and the OS is Android. In total, we recorded independent interaction data, in which about gesture trajectories were considered valid. After the experiments, the game experts conducted an interview with the participants to collect their subjective feedback when playing the mobile game app.
Virtual Joystick UI Verification. The game experts first focused on the triggering area and moving region of the joystick on the left side of the mobile screen. We applied interactive clustering on all the touching sampling dots generated by the participants’ interactions. To obtain the minimum coverage of the triggering area of the virtual joystick, we covered the interaction area with a yellow circle with the radius of . Through the interactive clustering, a new radius is determined. To obtain the upper boundary of the triggering area, we widened the initial coverage and adjusted the confidence coefficient to 99%. Another new radius was observed. Meanwhile, the distance to the left boundary and the bottom boundary is recorded to locate the specific position of the new center. The experimental results indicate that the lower and upper boundary of the triggering area is mm and mm based on our method and the distances to the left and the bottom boundary of the screen were mm and mm with the confidence coefficient of . The same procedure can be applied to determine the moving range of the joystick. We summarize our findings in Table 2 and Figure 8.

To further identify the major trajectories using the virtual joystick, we clustered the gestures that lie within the boundary of the moving area of the virtual joystick based on the proposed clustering method. Eexperimental results show the patterns that occupy the majority of the trajectories. To be specific, apart from those gestures with short distances, the gestures with relatively longer distances are clustered into clusters, representing the main gestures that demand high workload, i.e., moving fingers with a relatively long distance. As shown in Figure 9, the new boundary of moving area of the virtual joystick has covered the most frequent gestures that take a relatively long distance for fingers to move, which indicates that the above experimental scale suggestions can support a free movement according to the participants’ operation habits.
Skill Set UI Verification. Following a similar procedure, the game experts determined the scale and position of the skill responding area by adopting interactive clustering that covers all the touching sampling dots in the skill set area. Table 3 gives a summary result of the experimental results for the skill set. Particularly, with the confidence coefficient range of and , the spacing among skill buttons is in the range of mm – mm. Due to the fact that players may easily confuse with the middle two skill buttons, the spacing should be relatively larger between c and d. Similarly, the distances between the centers of the skill buttons and the right/bottom boundary of the mobile screen are in the range of mm – mm and mm – mm, while the distances between the center of the Normal Attack and the right/bottom boundary are in the range of mm – mm and mm – mm, as indicated in Figure 10.

Confidence Coefficient: 90% | ||||||||||
skill set | original radius | sampling number | original center | original number | new center | new radius | new number | distance to the left | distance to the bottom | diameter |
a | 169 | 8908 | (156.415, 146.934) | 3183 | (149.757, 146.309) | 79.890 | 2864 | 9.66 mm | 9.44 mm | 5.15 mm |
b | 90 | 8908 | (94.726, 353.550) | 747 | (94.396, 354.230) | 63.702 | 672 | 6.09 mm | 22.85 mm | 4.11 mm |
c | 80 | 8908 | (246.772, 321.063) | 1263 | (247.711, 321.218) | 74.087 | 1136 | 15.98 mm | 20.72 mm | 4.78 mm |
d | 80 | 8908 | (341.600, 243.923) | 1341 | (339.483, 247.096) | 71.902 | 1206 | 21.90 mm | 15.94 mm | 4.64 mm |
e | 80 | 8908 | (375.057, 92.975) | 1075 | (372.814, 91.963) | 71.109 | 967 | 24.05 mm | 5.93 mm | 4.59 mm |
Confidence Coefficient: 95% | ||||||||||
skill set | original radius | sampling number | original center | original number | new center | new radius | new number | distance to the left | distance to the bottom | diameter |
a | 169 | 8908 | (156.415, 146.934) | 3183 | (150.972, 145.146) | 109.730 | 3023 | 9.74 mm | 9.36 mm | 7.08 mm |
b | 90 | 8908 | (94.726, 353.550) | 747 | (94.782, 354.228) | 71.214 | 709 | 6.11 mm | 22.85 mm | 4.59 mm |
c | 80 | 8908 | (246.772, 321.063) | 1263 | (248.461, 320.519) | 77.606 | 1199 | 16.03 mm | 20.68 mm | 5.01 mm |
d | 80 | 8908 | (341.600, 243.923) | 1341 | (339.952, 246.550) | 88.189 | 1273 | 21.93 mm | 15.91 mm | 5.69 mm |
e | 80 | 8908 | (375.057, 92.975) | 1075 | (373.580,92.420) | 78.870 | 1021 | 24.10 mm | 5.96 mm | 5.09 mm |

6. Discussion and Reflections
6.1. System Performance and Generality
We first asked the game experts to evaluate the insights identified by our visualization approach. E.1 – 2 reported that “the visualization displays the interaction patterns of players intuitively”. Conventionally, the game team needed to watch video replays, and manually marked segments of interactions: “the time we spent on the entire process was about minutes since sometimes we had to replay a certain video segment many times,” said E.1. Our method can visualize the spatiotemporal attributes of the gesture interactions, making it easy for the experts to interact with players’ behaviors with the mobile game app instead of going through the entire video replay, which largely shortens the analysis time (now around minutes for each session). The UI design experts were very satisfied with our approach’s ability to spot potential UI design issues for the joystick and skill areas, which serves as a complementary to traditional methods such as survey other game competitors, “it helps me obtain detailed UI suggestions by interacting the interface,” said E.4, “I am now more confident to draw my UI design conclusions since my subjective feelings can be confirmed,” said E.5.
We also discussed with the five game experts which component(s) of our approach can be directly applied to other kinds of games or apps and which one(s) need customization. They all appreciated that our approach has been already applicable to any kind of ARPG apps since their functions are quite unified, mainly around joysticks and skills. The only place that needs customization is the middle part of the interface, i.e., the mobile game app interface that provides gaming contexts.
6.2. Design Implications for Mobile Game App UI
Our visualizations have provided some design implications for the UI design of all kinds of ARPG apps. Players spend most of the time operating with the joystick, which is also the most energy-consuming; therefore, it has the highest requirements for flexibility and fineness. “The purpose of the virtual joystick is to control the distance and direction of movement and the orientation of the game avatar,” said E.4. Participants also reported their requirements for the virtual joysticks include: 1) flexible and fast control of the position of the game avatar; 2) precise control of the distance and direction of movement; 3) lasting operation does not consume too much energy. The results from our visualizations also indicate that the finger movement areas should be sufficient enough to ensure that most areas of the mobile screen are responsive, “it is best to move anywhere you want,” said one participant (P1, male, age: 25).
In terms of the skill interaction design, UI design experts reported that their design principles are to support clicking on the graphical representations of the buttons to release skills and to display cooldowns when the skills are not available. From the perspective of participants, they wanted the design of the skill buttons to meet the following requirements: 1) to release skills quickly and easily; 2) to accurately release skills with less attention; 3) to quickly know when skills will be available. Through the quantitative experimental analysis of participants’ interaction data, UI design experts found that participants are easy to locate high-frequency buttons through the edge of the screen, which should be located along the “fan-shaped” curve. Other lower-frequency operation buttons such as switching roles can be placed in the area near the mm arc, facilitating easy clicking and locating. In addition, the participants reported that the middle two skill buttons are more easily noticed during operation and are suitable for major/important skills, while skills near the border on both sides are relatively more suitable for auxiliary skills such as dodging.
Our game experts also pointed out several directions for improving our visualization approach. E.1 and E.2 hoped that we could take in the in-game video and metrics from the mobile game app as a whole, enabling a comprehensive analysis. It is for ensuring the high “consistency of analysis conclusion.” Meanwhile, the game experts commented that “the method has the potential to be developed into a training tool to support a retrospective analysis of players’ performance.” The UI design experts (E.4 – 5) also hoped that we could include more in-game contents. For example, a proper number of monsters and stable in-game fighting duration together with appropriate UI design can not only prevent the players from “being in a state of high operating frequency all the time”, but also “bring the players a sense of satisfaction”.
7. Conclusion
In this work, we introduce a visualization approach to explore the interaction data of players on multi-touch mobile devices. The interaction data is transformed to gesture-based data and new visualization techniques are integrated to ease the exploration of interaction data. Three case studies and feedback from the game experts confirm the efficacy of our approach. In the future, we plan to include longitudinal studies to validate our approach and include other kinds of mobile game applications to better spot the limitations. Furthermore, embedding the game video in our visualization approach is a promising way to better learn how players behave.
Acknowledgements.
We are grateful for the valuable feedback and comments provided by the anonymous reviewers. This work is partially supported by the research start-up fund of ShanghaiTech University and HKUST-WeBank Joint Laboratory Project Grant No.: WEB19EG01-d.References
- (1)
- Anthony et al. (2015) Lisa Anthony, Quincy Brown, Jaye Nias, and Berthel Tate. 2015. Children (and adults) benefit from visual feedback during gesture interaction on mobile touchscreen devices. International Journal of Child-Computer Interaction 6 (2015), 17–27.
- Anthony et al. (2012) Lisa Anthony, Quincy Brown, Jaye Nias, Berthel Tate, and Shreya Mohan. 2012. Interaction and recognition challenges in interpreting children’s touch and gesture input on mobile devices. In Proceedings of the 2012 ACM International Conference on Interactive Tabletops and Surfaces. 225–234.
- Anthony et al. (2013) Lisa Anthony, Radu-Daniel Vatavu, and Jacob O. Wobbrock. 2013. Understanding the Consistency of Users’ Pen and Finger Stroke Gesture Articulation. In Proceedings of Graphics Interface 2013 (Regina, Sascatchewan, Canada) (GI ’13). Canadian Information Processing Society, CAN, 87–94.
- Anthony (2012) T Anthony. 2012. Finger-friendly design: ideal mobile touchscreen target sizes. Accessed 27, 4 (2012), 2014.
- Baldauf et al. (2015) Matthias Baldauf, Peter Fröhlich, Florence Adegeye, and Stefan Suette. 2015. Investigating on-screen gamepad designs for smartphone-controlled video games. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM) 12, 1s (2015), 1–21.
- Banovic et al. (2019) Nikola Banovic, Antti Oulasvirta, and Per Ola Kristensson. 2019. Computational modeling in human-computer interaction. In Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems. 1–7.
- Benko et al. (2006) Hrvoje Benko, Andrew D Wilson, and Patrick Baudisch. 2006. Precise selection techniques for multi-touch screens. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 1263–1272.
- Blascheck et al. (2014) Tanja Blascheck, Kuno Kurzhals, Michael Raschke, Michael Burch, Daniel Weiskopf, and Thomas Ertl. 2014. State-of-the-Art of Visualization for Eye Tracking Data.. In EuroVis (STARs).
- Bragdon et al. (2011) Andrew Bragdon, Eugene Nelson, Yang Li, and Ken Hinckley. 2011. Experimental analysis of touch-screen gesture designs in mobile environments. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 403–412.
- Bunian et al. (2017) Sara Bunian, Alessandro Canossa, Randy Colvin, and Magy Seif El-Nasr. 2017. Modeling individual differences in game behavior using HMM. In Proceedings of the 13th Artificial Intelligence and Interactive Digital Entertainment Conference.
- Burzacca and Paternò (2013) Paolo Burzacca and Fabio Paternò. 2013. Analysis and visualization of interactions with mobile web applications. In Proceedings of IFIP Conference on Human-Computer Interaction. Springer, 515–522.
- Carta et al. (2011) Tonio Carta, Fabio Paternò, and Vagner Santana. 2011. Support for remote usability evaluation of web mobile applications. In Proceedings of the 29th ACM International Conference on Design of Communication. 129–136.
- Clark (2010) Josh Clark. 2010. Tapworthy: Designing great iPhone apps. ” O’Reilly Media, Inc.”.
- Folmer (2007) Eelke Folmer. 2007. Designing usable and accessible games with interaction design patterns. Retrieved April 10 (2007), 2008.
- Froehlich et al. (2007) Jon Froehlich, Mike Y Chen, Sunny Consolvo, Beverly Harrison, and James A Landay. 2007. MyExperience: a system for in situ tracing and capturing of user feedback on mobile phones. In Proceedings of the 5th International Conference on Mobile Systems, Applications and Services. 57–70.
- Guo et al. (2013) Qi Guo, Haojian Jin, Dmitry Lagun, Shuai Yuan, and Eugene Agichtein. 2013. Mining touch interaction data on mobile devices to predict web search result relevance. In Proceedings of the 36th International ACM SIGIR Conference on Research and Development in Information Retrieval. 153–162.
- Harada et al. (2013) Susumu Harada, Daisuke Sato, Hironobu Takagi, and Chieko Asakawa. 2013. Characteristics of elderly user behavior on mobile multi-touch devices. In Proceedings of IFIP Conference on Human-Computer Interaction. Springer, 323–341.
- Harty et al. (2021) Julian Harty, Haonan Zhang, Lili Wei, Luca Pascarella, Mauricio Aniche, and Weiyi Shang. 2021. Logging Practices with Mobile Analytics: An Empirical Study on Firebase. arXiv preprint arXiv:2104.02513 (2021).
- Hellstén (2019) Tuomas Hellstén. 2019. Iterative design of mobile game UX: Design defined by a target group. (2019).
- Hesenius et al. (2014) Marc Hesenius, Tobias Griebe, Stefan Gries, and Volker Gruhn. 2014. Automating UI tests for mobile applications with formal gesture descriptions. In Proceedings of the 16th International Conference on Human-Computer Interaction with Mobile Devices & Services. 213–222.
- Huang et al. (2012) Jeff Huang, Ryen White, and Georg Buscher. 2012. User see, user point: gaze and cursor alignment in web search. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 1341–1350.
- Humayoun et al. (2017) Shah Rukh Humayoun, Paresh Hamirbhai Chotala, Muhammad Salman Bashir, and Achim Ebert. 2017. Heuristics for evaluating multi-touch gestures in mobile applications. In Proceedings of the 31st International BCS Human Computer Interaction Conference (HCI 2017) 31. 1–6.
- Kang et al. (2013) Ah Reum Kang, Jiyoung Woo, Juyong Park, and Huy Kang Kim. 2013. Online game bot detection based on party-play log analysis. Computers & Mathematics with Applications 65, 9 (2013), 1384–1395.
- Kim et al. (2015) Jaewon Kim, Paul Thomas, Ramesh Sankaranarayana, Tom Gedeon, and Hwan-Jin Yoon. 2015. Eye-tracking analysis of user behavior and performance in web search on large and small screens. Journal of the Association for Information Science and Technology 66, 3 (2015), 526–544.
- Korhonen and Koivisto (2006) Hannu Korhonen and Elina MI Koivisto. 2006. Playability heuristics for mobile games. In Proceedings of the 8th Conference on Human-Computer Interaction with Mobile Devices and Services. 9–16.
- Krieter and Breiter (2018) Philipp Krieter and Andreas Breiter. 2018. Analyzing mobile application usage: generating log files from mobile screen recordings. In Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services. 1–10.
- Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems 25 (2012), 1097–1105.
- Lee et al. (2014) Seong Jae Lee, Yun-En Liu, and Zoran Popovic. 2014. Learning individual behavior in an educational game: a data-driven approach. In Educational Data Mining 2014. Citeseer.
- Leiva et al. (2018) Luis A. Leiva, Daniel Martín-Albo, and Radu-Daniel Vatavu. 2018. GATO: Predicting Human Performance with Multistroke and Multitouch Gesture Input. In Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services (Barcelona, Spain) (MobileHCI ’18). Association for Computing Machinery, New York, NY, USA, Article 32, 11 pages. https://doi.org/10.1145/3229434.3229478
- Li et al. (2018b) Quan Li, Kristanto Sean Njotoprawiro, Hammad Haleem, Qiaoan Chen, Chris Yi, and Xiaojuan Ma. 2018b. EmbeddingVis: A Visual Analytics Approach to Comparative Network Embedding Inspection. In Proceedings of 2018 IEEE Conference on Visual Analytics Science and Technology (VAST). 48–59. https://doi.org/10.1109/VAST.2018.8802454
- Li et al. (2016) Quan Li, Peng Xu, Yeuk Yin Chan, Yun Wang, Zhipeng Wang, Huamin Qu, and Xiaojuan Ma. 2016. A visual analytics approach for understanding reasons behind snowballing and comeback in moba games. IEEE transactions on visualization and computer graphics 23, 1 (2016), 211–220.
- Li et al. (2019) Wei Li, Mathias Funk, Quan Li, and Aarnout Brombacher. 2019. Visualizing event sequence game data to understand player’s skill growth through behavior complexity. Journal of Visualization 22, 4 (2019), 833–850.
- Li (2010) Yang Li. 2010. Protractor: A Fast and Accurate Gesture Recognizer. Association for Computing Machinery, New York, NY, USA, 2169–2172. https://doi.org/10.1145/1753326.1753654
- Li et al. (2018a) Yang Li, Samy Bengio, and Gilles Bailly. 2018a. Predicting Human Performance in Vertical Menu Selection Using Deep Learning. Association for Computing Machinery, New York, NY, USA, 1–7. https://doi.org/10.1145/3173574.3173603
- Lin et al. (2018) Yu-Hao Lin, Suwen Zhu, Yu-Jung Ko, Wenzhe Cui, and Xiaojun Bi. 2018. Why Is Gesture Typing Promising for Older Adults? Comparing Gesture and Tap Typing Behavior of Older with Young Adults. In Proceedings of the 20th International ACM SIGACCESS Conference on Computers and Accessibility (Galway, Ireland) (ASSETS ’18). Association for Computing Machinery, New York, NY, USA, 271–281. https://doi.org/10.1145/3234695.3236350
- Miniukovich and De Angeli (2015) Aliaksei Miniukovich and Antonella De Angeli. 2015. Computation of interface aesthetics. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. 1163–1172.
- Nebeling et al. (2013) Michael Nebeling, Maximilian Speicher, and Moira Norrie. 2013. W3touch: metrics-based web page adaptation for touch. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 2311–2320.
- Ouyang and Li (2012) Tom Ouyang and Yang Li. 2012. Bootstrapping Personal Gesture Shortcuts with the Wisdom of the Crowd and Handwriting Recognition. Association for Computing Machinery, New York, NY, USA, 2895–2904. https://doi.org/10.1145/2207676.2208695
- Park et al. (2013) Jaehyun Park, Sung H Han, Hyun K Kim, Youngseok Cho, and Wonkyu Park. 2013. Developing elements of user experience for mobile phones and services: survey, interview, and observation approaches. Human Factors and Ergonomics in Manufacturing & Service Industries 23, 4 (2013), 279–293.
- Punchoojit and Hongwarittorrn (2017) Lumpapun Punchoojit and Nuttanont Hongwarittorrn. 2017. Usability studies on mobile user interface design patterns: a systematic literature review. Advances in Human-Computer Interaction 2017 (2017).
- Ruiz et al. (2011) Jaime Ruiz, Yang Li, and Edward Lank. 2011. User-defined motion gestures for mobile interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 197–206.
- Swearngin and Li (2019) A. Swearngin and Y. Li. 2019. Modeling Mobile Interface Tappability Using Crowdsourcing and Deep Learning. (2019).
- Tu et al. (2015) Huawei Tu, Xiangshi Ren, and Shumin Zhai. 2015. Differences and Similarities between Finger and Pen Stroke Gestures on Stationary and Mobile Devices. ACM Trans. Comput.-Hum. Interact. 22, 5, Article 22 (Aug. 2015), 39 pages. https://doi.org/10.1145/2797138
- Vatavu and Ungurean (2019) Radu-Daniel Vatavu and Ovidiu-Ciprian Ungurean. 2019. Stroke-Gesture Input for People with Motor Impairments: Empirical Results & Research Roadmap. Association for Computing Machinery, New York, NY, USA, 1–14. https://doi.org/10.1145/3290605.3300445
- Wobbrock et al. (2007) Jacob O. Wobbrock, Andrew D. Wilson, and Yang Li. 2007. Gestures without Libraries, Toolkits or Training: A $1 Recognizer for User Interface Prototypes. In Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology (Newport, Rhode Island, USA) (UIST ’07). Association for Computing Machinery, New York, NY, USA, 159––168. https://doi.org/10.1145/1294211.1294238
- Wu et al. (2020) Ziming Wu, Yulun Jiang, Yiding Liu, and Xiaojuan Ma. 2020. Predicting and diagnosing user engagement with mobile ui animation via a data-driven approach. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1–13.
- Wu et al. (2019) Ziming Wu, Taewook Kim, Quan Li, and Xiaojuan Ma. 2019. Understanding and modeling user-perceived brand personality from mobile application uis. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–12.
- Zhai et al. (2012) Shumin Zhai, Per Ola Kristensson, Caroline Appert, Tue Haste Andersen, and Xiang Cao. 2012. Foundational issues in touch-screen stroke gesture design-an integrative review. Foundations and Trends in Human-Computer Interaction 5, 2 (2012), 97–205.