Paper deep dive
Optimizing Feature Extraction for On-device Model Inference with User Behavior Sequences
Chen Gong, Zhenzhe Zheng, Yiliu Chen, Sheng Wang, Fan Wu, Guihai Chen
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 95%
Last extracted: 3/26/2026, 2:29:20 AM
Summary
AutoFeature is an automated feature extraction engine designed to optimize on-device machine learning model execution by identifying and eliminating redundant operations across different input features and consecutive model inferences. It utilizes graph abstraction, graph optimization, and efficient caching to reduce latency in mobile applications without compromising model accuracy.
Entities (4)
Relation Signals (3)
AutoFeature → optimizes → Feature Extraction
confidence 100% · AutoFeature, an automated feature extraction engine designed to accelerate on-device feature extraction process
AutoFeature → uses → FE-graph
confidence 95% · AutoFeature represents its feature extraction process as a directed acyclic graph, termed FE-graph
Feature Extraction → isbottleneckfor → On-device Machine Learning
confidence 90% · we identify an overlooked bottleneck in real-world on-device model execution pipelines: extracting input features
Cypher Suggestions (2)
Map the relationship between the system and the process it optimizes · confidence 95% · unvalidated
MATCH (s:System {name: 'AutoFeature'})-[r:OPTIMIZES]->(p:Process) RETURN s.name, type(r), p.nameFind all optimization techniques used by AutoFeature · confidence 90% · unvalidated
MATCH (s:System {name: 'AutoFeature'})-[:USES]->(t:Technique) RETURN t.nameAbstract
Abstract:Machine learning models are widely integrated into modern mobile apps to analyze user behaviors and deliver personalized services. Ensuring low-latency on-device model execution is critical for maintaining high-quality user experiences. While prior research has primarily focused on accelerating model inference with given input features, we identify an overlooked bottleneck in real-world on-device model execution pipelines: extracting input features from raw application logs. In this work, we explore a new direction of feature extraction optimization by analyzing and eliminating redundant extraction operations across different model features and consecutive model inferences. We then introduce AutoFeature, an automated feature extraction engine designed to accelerate on-device feature extraction process without compromising model inference accuracy. AutoFeature comprises three core designs: (1) graph abstraction to formulate the extraction workflows of different input features as one directed acyclic graph, (2) graph optimization to identify and fuse redundant operation nodes across different features within the graph; (3) efficient caching to minimize operations on overlapping raw data between consecutive model inferences. We implement a system prototype of AutoFeature and integrate it into five industrial mobile services spanning search, video and e-commerce domains. Online evaluations show that AutoFeature reduces end-to-end on-device model execution latency by 1.33x-3.93x during daytime and 1.43x-4.53x at night.
Tags
Links
- Source: https://arxiv.org/abs/2603.21508v1
- Canonical: https://arxiv.org/abs/2603.21508v1
Full Text
95,662 characters extracted from source content.
Expand or collapse full text
Optimizing Feature Extraction for On-device Model Inference with User Behavior Sequences Chen Gong gongchen@sjtu.edu.cn Shanghai Jiao Tong University Shanghai, China Zhenzhe Zheng ∗ zhengzhenzhe@sjtu.edu.cn Shanghai Jiao Tong University Shanghai, China Yiliu Chen chenyiliu@bytedance.com ByteDance Hangzhou, China Sheng Wang wangsheng.john@bytedance.com ByteDance Hangzhou, China Fan Wu fwu@cs.sjtu.edu.cn Shanghai Jiao Tong University Shanghai, China Guihai Chen gchen@cs.sjtu.edu.cn Shanghai Jiao Tong University Shanghai, China Abstract Machine learning models are widely integrated into modern mo- bile apps to analyze user behaviors and deliver personalized ser- vices. Ensuring low-latency on-device model execution is critical for maintaining high-quality user experiences. While prior research has primarily focused on accelerating model inference with given input features, we identify an overlooked bottleneck in real-world on-device model execution pipelines: extracting input features from raw application logs. In this work, we explore a new direction of feature extraction optimization by analyzing and eliminating re- dundant extraction operations across different model features and consecutive model inferences. We then introduce AutoFeature, an automated feature extraction engine designed to accelerate on- device feature extraction process without compromising model inference accuracy. AutoFeature comprises three core designs: (1) graph abstraction to formulate the extraction workflows of different input features as one directed acyclic graph, (2) graph optimization to identify and fuse redundant operation nodes across different fea- tures within the graph; (3) efficient caching to minimize operations on overlapping raw data between consecutive model inferences. We implement a system prototype of AutoFeature and integrate it into five industrial mobile services spanning search, video and e-commerce domains. Online evaluations show that AutoFeature reduces end-to-end on-device model execution latency by 1.33×- 3.93× during daytime and 1.43×-4.53× at night. CCS Concepts • Human-centered computing→Mobile computing;• Com- puting methodologies→ Machine learning. Keywords On-Device Machine Learning; User Behavior Analysis; Model In- ference Acceleration; Feature Extraction Optimization ∗ Zhenzhe Zheng is the corresponding author. This work is licensed under a Creative Commons Attribution-NonCommercial- NoDerivatives 4.0 International License. SenSys ’26, Saint Malo, France © 2026 Copyright held by the owner/author(s). ACM ISBN 979-8-4007-2309-4/2026/05 https://doi.org/10.1145/3774906.3800493 ACM Reference Format: Chen Gong, Zhenzhe Zheng, Yiliu Chen, Sheng Wang, Fan Wu, and Guihai Chen. 2026. Optimizing Feature Extraction for On-device Model Inference with User Behavior Sequences. In ACM/IEEE International Conference on Embedded Artificial Intelligence and Sensing Systems (SenSys ’26), May 11– 14, 2026, Saint Malo, France. ACM, New York, NY, USA, 15 pages. https: //doi.org/10.1145/3774906.3800493 1 Introduction With the rapid advancements of mobile devices, machine learning (ML) models are increasingly integrated into modern mobile apps to deliver personalized services [4,22,24,43,61,75]. Unlike traditional large-scale vision or language models that use static input features (e.g., image pixels or tokens embeddings), real-world ML models deployed on mobile devices for industrial apps are typically smaller and rely on dynamic input features extracted from evolving user behavior sequences to capture user intent (e.g., genre list of the last 5 watched videos reflect shifting interests). Typical examples include customized product advertisements on e-commerce platforms [13, 14], video recommendation and preloading in video apps [50,67,76] and search result ranking in search engines [35, 80, 81]. For on-device deployment, ensuring low-latency model execu- tion is essential for maintaining both service quality and user expe- rience. In real-world mobile service workflows as shown in Figure 1, the device first extracts input features from historical user be- haviors recorded in application logs (app log), and then performs model inference to predict user’s current intent, with the ultimate goal of generating personalized app responses (e.g., preload suitable next-to-watch videos). High latency of on-device model execution not only blocks follow-up app responses at the system level, but also causes stuttering user experiences. Empirical studies show that even a slight 0.15% increase in stuttering rate can lead to the loss of 9 million app users [1]. Therefore, on-device ML model execution is expected to be imperceptible to users, ideally within 30 ms to match the human perception range of 30-60 FPS [16]. Feature Extraction Bottleneck. For vision and language mod- els, significant efforts have been made to optimize the efficiency of on-device model inference stage from both algorithm and hardware aspects [23,25,28,31,33,36,38,41,51,65,69,77]. However, our analysis of on-device model execution pipelines in real-world mo- bile apps reveals an overlooked bottleneck: extracting user features from raw app logs accounts for 61-86% of the total model execution arXiv:2603.21508v1 [cs.LG] 23 Mar 2026 SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. Inference Request Mobile App Service App Log Input Features Predictions Bandwidth Control Recomm- endation Content Preloading Resolution Adjust Application Responses On-Device Model Execution Feature Extraction Model Inference Figure 1: Workflow of real-world mobile application services. latency. This issue stems from three key factors (elaborated in §2.2): (i) A large proportion of input features (74% on average) required by on-device models for mobile apps are user features to reflect various user behaviors; (i) Extracting each user feature involves multiple resource-intensive operations on raw behavior data in app logs; (i) The on-device model inference is relatively fast due to model size limits and mature optimization techniques. Motivation. In this work, we delve into a crucial but unexplored direction, feature extraction optimization, to accelerate on-device model execution for mobile apps without sacrificing accuracy. Our core insight is that many input features required by on-device models are extracted from overlapping user behavior data in app logs (§2.3). This implies that redundant operations exist in extract- ing (i) different input features within each model execution, and (i) identical input features across consecutive model executions. These redundancies motivate us to alleviate feature extraction bottleneck by eliminating unnecessary data processing operations. Overall Design. We introduce AutoFeature, an efficient feature extraction engine for faster on-device model execution by elimi- nating redundant operations across different input features and consecutive model executions. For an ML model deployed in the mobile app, AutoFeature represents its feature extraction process as a directed acyclic graph, termed FE-graph, where source node de- notes raw app log data, each target node denotes a feature and they are connected by a chain of operation nodes. Next, AutoFeature optimizes the FE-graph from two aspects. (i) Inter-feature: within a single model execution, AutoFeature identifies and fuses redun- dant operation nodes across different features in the FE-graph; (i) Cross-execution: across consecutive inference requests, AutoFea- ture reuses intermediate results from previous model executions to eliminate redundant operations on overlapping data. Since AutoFea- ture is designed to work independently before the model inference stage, it can be seamlessly integrated with any device operating systems, back-end mobile inference engines and ML models de- veloped by different teams within a mobile app enterprise. It also complements previous efforts on model inference acceleration. Challenges and Our Solutions. AutoFeature addresses three major challenges in formulating and optimizing the FE-graph for single and multiple model executions. First, constructing a unified FE-graph for automatic redundancy identification is non-trivial. Extracting input features from raw app logs is a complex process and existing literature lacks systematic analysis of it. Such opacity complicates the abstraction of feature extraction process into discrete, critical operation nodes that could facilitate automated redundancy detection across features. To solve this, AutoFeature characterizes feature extraction as an information filtering process, which leverages multiple orthogonal conditions to progressively filter necessary information from raw app log data. Each feature’s extraction process is then abstracted as a chain of operation nodes, each associated with different filtering conditions. In this way, any inter-feature redundancy can be systematically quantified by computing the intersections of filtering conditions of their operation nodes (§3.2). Second, efficiently eliminating redundancy across features is chal- lenging. Intuitively, redundancies between any features with over- lapping conditions can be eliminated by fusing their chains of operation nodes within the FE-graph. However, without careful designs on how to start and terminate the chain fusion process, the acceleration benefits can be offset by the introduced costs, in- cluding: (i) operations on irrelevant app log data, as the fused fil- tering range of multiple conditions can be broader than their orig- inally intended condition scope, and (i) extra termination costs to separate the node outputs for fused features. To address this, instead of treating each feature’s operation chain as a monolithic unit, AutoFeature decomposes it into multiple sub-chains with narrower condition per node, exposing finer-grained node fusion opportunities without condition scope expansion. Further, a hier- archical filtering algorithm is proposed to progressively separate outputs for fused features based on their condition relations, re- ducing termination cost from푂 푙푒푛(표푢푡푝푢푡푠)×푛푢푚(푓푒푎푡푢푟푒푠) to 푂 푙푒푛(표푢푡푝푢푡푠)+푛푢푚(푓푒푎푡푢푟푒푠) . Third, minimizing redundant operations across consecutive model executions is not straightforward either. Although caching all in- termediate results for each model execution could eliminate inter- execution redundancy, this approach is not always feasible and could cause potential app crashes due to the dynamic and limited memory space allocated to each ML model of each mobile app. To optimally balance redundancy elimination and memory cost, AutoFeature formulates the caching decision as a classic knapsack packing problem, where the object is to maximize computational savings within a given memory budget. A greedy policy is then proposed to prioritize intermediate results with the highest benefit- to-cost ratios, which can be efficiently measured in constant time complexity through our term decomposition technique (§3.4). Contributions of this work are summarized as follows: •To the best of our knowledge, we are the first to unveil and an- alyze the feature extraction bottleneck in practical on-device ML model executions, exploring feature extraction optimization as a new research problem. •We propose AutoFeature system to address feature extraction bottleneck without compromising model accuracy by eliminating redundant operations across different input features and consecu- tive model executions. •We demonstrate AutoFeature’s remarkable performance through extensive online evaluations in real-world mobile services, covering domains of search, video and e-commerce. 2 Background and Motivation In this section, we first elaborate on-device model execution pipelines in mobile apps (§2.1). Then, we analyze feature extraction bottle- neck in industrial mobile service workloads (§2.2) and explore the optimization opportunities that motivate our work (§2.3). Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France Timestamp = 000010 Event_name = “video_play” Duration = 300 Genre = “Comedy” Window_Ratio = 0.5 . . . Capture . . . User Behaviors Behavior EventApplication Log (SQLite Database) “Timestamp”: int(000010), “Event_name”: str(video_play), “Duration”: float(300), “Genre”: str(Comedy), “Window_Ratio”: float(0.5), ... Relevent Event Entries Decoded Attributes Necessary Attributes Features Outputs Forward Record Filter Compute Decode ① Behavior Logging ② Feature Extraction ③ Model Inference TimestampEvent NameCompressed Behavior-Specific Attributes 000010 video_playJSON(‘“Duration”:300, “Genre”:”Comedy”, ...’) 000020 shareJSON(‘“Author_ID”:3321, “Content”:”Article”, ...’) ... ...... 000010 video_playJSON(‘“Duration”:300, “Genre”: ...’) 000050 video_playJSON(‘“Duration”:15, “Genre”: ...’) 001010 video_playJSON(‘“Duration”:600, “Genre”: ...’) 300 15 600 . . . User Device Cloud Retrieve Prior Efforts Our Focus Video Play Search Add to Cart ... Figure 2: A complete on-device model execution pipeline in industrial mobile apps. 2.1 On-Device Model Execution Pipeline As shown in Figure 2, a complete on-device model execution pipeline in industrial mobile apps consists of three key stages that transform physical user behaviors into input features usable for ML models and ultimately into user intent predictions. Stage 1: Behavior Logging. In mobile apps, each interaction behavior between user and the graphical user interface (GUI) can be captured as a behavior event (e.g., Search, Add-to-Cart, Video-Play). Each behavior event is a structured format of multiple attributes, including behavior-independent attributes (e.g., timestamp, event name) and behavior-specific attributes that vary by behavior type to provide detailed descriptions (e.g., duration and genre of Video- Play, item_id and price of Add_to_Cart). Then, each behavior event is recorded as a single row in app log, which is typically managed using SQLite database [25,39,55] on both iOS and Android de- vices [20,53]. In Figure 3, our analysis of 100 common behavior types from a popular video app shows that 50% of user behavior types contain more than 25 attributes and 25% contain over 85 attributes. To manage this complexity and reduce storage costs, for each row of behavior event, behavior-independent attributes are stored in separate columns for data retrieval, while behavior- specific attributes are typically compressed into a single column 1 , as shown in the gray part of Figure 2. Stage 2: Feature Extraction. When an on-device model exe- cution is invoked by a mobile service, the device extracts all input features required by the ML model to fully reflect the ongoing user context, including: (i) user features to summarize various user behav- ior types over different time periods (e.g., average duration of videos watched over the past hour or day), (i) device features to describe the current device state (e.g., volume level, battery percent), and (i) cloud features to supplement information from service providers (e.g., embeddings of user_id and service_id). While device and cloud features are either readily accessible or pre-fetched from cloud in advance, the specific values of user features are dynamic with time and require real-time extraction from the latest app logs. Stage 3: Model Inference. Once all required features are ex- tracted, the on-device model performs inference to produce pre- dictions for subsequent system responses. This process has been systematically supported by well-established mobile ML engines such as TensorFlow Lite [2], MNN [33] and ByteNN [54], which are optimized for efficient execution on mobile hardware. 1 Storing behavior-specific attributes in separate columns would lead to excessive null values in app log and high storage cost [25], as different behaviors have heterogeneous attributes for description. 04080120160 # of Behavior-Specific Attributes 0.00 0.25 0.50 0.75 1.00 CDF across Behaviors (25, 0.5) (85, 0.75) Figure 3: Attribute number of mobile user behaviors. S1S2S3S4V1V2V3V4E1E2 Workloads (S:Search V:Video E:E-Commerce) 0 20 40 60 80 100 Latency Breakdown (%) Feature ExtractionModel Inference Figure 4: Time breakdown of on-device model execution. 0.000.250.500.751.00 Proportion of User Features 0.00 0.25 0.50 0.75 1.00 CDF across Models (0.73, 0.5) 04080120160 Number of Different Features 0.00 0.25 0.50 0.75 1.00 CDF across Models Cloud Feature Device Feature User Feature Total Feature Figure 5: Proportion of user features (left) and numbers of various features (right) across on-device ML models. 2.2 Feature Extraction Bottleneck To achieve low-latency on-device model execution, prior academic research has predominantly focused on optimizing the model infer- ence stage (see §6), as they targeted traditional vision and language models that use static input features like image pixels and word embeddings. In Figure 4, we observe real-world on-device model execution pipelines in industrial mobile apps across domains of search, video and e-commerce (§4.1) and break down their latencies. Our empirical study unveils an overlooked performance bottleneck: feature extraction alone accounts for 61%-86% of the total end-to-end model execution latency. This significant bottleneck stems from three primary factors. High Proportion of User Features. For popular mobile apps like TikTok and Taobao, the ML models practically deployed on devices are used to analyze user intent based on private and personal user behavior data. As a result, these models rely heavily on user features extracted from various behavior types across different time windows. We analyze the proportions and numbers of user features across over 20 ML models deployed in mobile apps that collaborate with us and present the results in Figure 5. We notice that on average, user features constitute 73% of the total input features required by ML models. Specifically, 50% of on-device ML SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. 020406080100120 ID of User Feature (sorted by event name and time range) 60s 10m 1h 4h 24h Time Range of Features click_A close_A comment_A dislike_A enter_A enter_B enter_C enter_D enter_E enter_F enter_G favourite_A follow_A homepage_A inner_A like_A pitaya_A play_A post_A share_A share_B video_A video_B video_C (a) Features can require the same behavior type with overlapped time ranges. 10s1min30min1h12h1day Model Inference Interval 0.00 0.25 0.50 0.75 1.00 Redundancy Time Ranges of Features 5min1hour1day1week 0.20.40.60.81.0 Redundancy between Inferences 0.00 0.25 0.50 0.75 1.00 CDF across Models (0.34, 0.25) (0.36, 0.5) (0.43, 0.75) (0.99, 1.0) (b) Cross-inference redundancy of different features (left) and models (right). Figure 6: Analysis on real-world on-device ML models to reveal redundancy across features and inferences. models require more than 60 distinct user features and 20% require as many as 110 user features. Cumbersome User Feature Extraction. Extracting user fea- tures from raw app logs is resource-intensive and time-consuming, due to the misaligned granularity between data stored in app logs and required by model features. As explained in §2.1 and visual- ized in Figure 2, user behaviors are recorded by mobile apps at event level in app logs, where each row corresponds to a specific behavior event and behavior-specific attributes are compressed in one column for efficient storage. However, each user feature typically correlates with only a few attributes of certain behav- ior types (i.e., attribute level), such as average watching durations, genre list of videos watched over past hour. This is because dif- ferent attributes are designed to describe the same behavior from diverse dimensions (e.g., watching duration, pause times, volume, genre, etc). As a result, extracting each user feature requires mul- tiple resource-intensive operations, including retrieving relevant event rows, decompressing detailed attributes and filtering neces- sary attributes for computation (elaborated in yellow part of Figure 2 and §3.2). Fast On-Device Model Inference. Compared with the high latency of feature extraction, the practical on-device model infer- ence is relatively fast, typically completed within millisecond-level latency, which is due to three root causes: (i) Model size limita- tion: Mobile platforms have strict limits on app size (e.g., 2GB for iOS [17] and 4GB for Android [26]), which limits the size of each ML model deployed by mobile app on native devices. (i) Simple task: Most ML models practically offloaded to mobile devices handle lightweight but real-time prediction tasks for subsequent system re- sponses, where compact models like decision trees [58], multi-layer perceptrons [59] and small neural networks [13] are qualified. (i) Advanced optimization: Years of research has led to highly efficient mobile inference engines [2,33], with the help of mature optimiza- tion techniques from hardware, operator, model and algorithm aspects [25, 28, 31, 33, 36, 38, 44, 51, 65, 69]. 2.3 Optimization Opportunities In this work, we aim to address feature extraction bottleneck for real-world on-device model execution by eliminating redundant operations on overlapping data across both different input features and successive model inferences. Inter-Feature Redundancy. After scrutinizing ML models de- ployed in industrial mobile apps, we observe that many user fea- tures required by the same model can rely on overlapping behavior events, resulting in repeated operations on the same event rows in app log. Specifically, in Figure 6(a), we visualize the behavior types and time periods required by input features of a video rec- ommendation model in TikTok, with feature names and behavior types anonymized for privacy. We observe that although the model requires 134 distinct user features, they are extracted from only 24 unique behavior types (indicated by colored bars and hatches). De- spite variations in their target time windows and behavior-specific attributes, raw event rows processed by different features can re- main largely overlapped. Such redundancy presents a substantial opportunity to optimize feature extraction process by fusing opera- tions on overlapping events necessitated by different features. Cross-Inference Redundancy. Redundant operations also oc- cur between consecutive model executions triggered by the same mobile service. In mobile apps, user intent and preferences are dynamic, requiring frequent model inferences to deliver real-time, accurate system responses. When the interval between model execu- tions is shorter than the time range required for feature extraction, event rows processed in previous execution remains relevant and reusable for next execution. Figure 6(b) illustrates this pattern. For features based on the last 5 minutes of behavior, 60% of relevant event rows can be reused when model inference is triggered every minute. This overlap increases to 90% for features that rely on be- havior within the last hour. We further show that this redundancy is widespread by collecting cross-inference redundancies of ML models running online in mobile apps. As shown in Figure 6(b), 75% on-device models exhibit over 34% overlapping data between online inferences and 25% exhibit overlap exceeding 43%. These findings highlight the inefficiencies in current feature extraction process, inspiring us to design a more efficient feature extraction engine for reducing on-device model execution latency without compromising model accuracy. 3 AutoFeature Design In this work, we introduce AutoFeature, a universal feature ex- traction engine designed to automatically identify and eliminate redundant operations across different features and consecutive model executions, accelerating end-to-end execution of on-device ML models for mobile apps. 3.1 Overview As depicted in Figure 7, AutoFeature serves as an independent opti- mization layer that integrates seamlessly with existing on-device model execution pipelines. It complements existing optimizations on model inference stage and operates in two phases: offline opti- mization and online execution. Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France Extract from New Events App Service Inference Request Mobile Device Model Inference Model Prediction ML Model Graph Generator Intra-Feature Partition Inter-Feature Fusion Output Evaluator Raw FE-GraphGranular FE-Graph Optimized FE-Graph Cache Policy 1 2 4 Features Online Execution LOG Offline Optimization Optimized FE-Graph Cache Fetch Past Results 3 1 4 Cache Policy ① User Feature ② ③ ③ ④ Figure 7: AutoFeature overview and workflow. Offline Optimization. When a new ML model is downloaded by mobile device via an app update, AutoFeature performs a one- time offline optimization to restructure the feature extraction work- flow, which includes three main components. First,❶graph gen- erator formulates the feature extraction workflow as a directed acyclic graph (FE-graph), where each source node of app log and each target node of an input feature is connected by a chain of critical operation nodes (§3.2). Second, AutoFeature optimizes the FE-graph to eliminate redundant operations across features (§3.3), including❷intra-feature partition to decompose each feature’s op- eration chain into smaller sub-chains to expose finer-grained fusion opportunities for redundant operation nodes, and❸inter-feature fusion to judiciously fuse operation nodes with overlapping inputs to eliminate redundancy and carefully design termination node to minimize output separation cost for fused node. Third,❹output evaluator valuates each node’s outputs based on the ratio of poten- tial computation savings to memory footprint. A greedy caching policy is then applied to select high-value intermediate results for reuse during online execution (§3.4). Online Execution. At run time, when an model execution re- quest is issued, AutoFeature accelerates feature extraction through the following steps:①fetching previously computed intermediate results from the cache,②extracting missing intermediate results for newly logged user behavior data using the optimized FE-graph, ③assembling cached and newly computed results to reconstruct real-time user features, and④updating intermediate results in cache based on the cache policy and memory constraints. 3.2 Graph Generator: Automated Redundancy Identification Given an on-device ML model, AutoFeature first needs to model the extraction process of each feature as a chain of discrete operation nodes, enabling systematic identification of redundancies across features. Such design is inspired by successful modern ML frame- works (e.g., TensorFlow, PyTorch, MNN), which compile models into computational graphs to enable operator fusion and sched- uling. Similarly, as a feature extraction engine, AutoFeature aims to abstract the entire feature extraction workflow into a graph for systematic redundancy identification and elimination. To achieve this, we characterize the feature extraction process as a form of information filtering [7], which progressively removes unwanted Compute (comp_func) count, min/max, concate Decode() Retrieve (event_names, time_range) t i m e s t a m p t i m e s t a m p t i m e s t a m p a t t r i b u t e s event_nameevent_name event_name a t t r i b u t e s t i m e s t a m p event_name Application LogRelevent EntriesDecoded Attributes Filtered Attribute ListUser Feature Filter (attr_name) Figure 8: Four atomic operations of feature extraction. information from an information stream using filtering conditions. Through an empirical analysis of user features required by on- device ML models, we conclude that any user feature can be defined by a set of four orthogonal conditions: < 푒푣푒푛푡 _푛푎푚푒푠,푡푖푚푒_푟푎푛푔푒,푎푡푟 _푛푎푚푒,푐표푚푝_푓푢푛푐>, where event_name specifies the behavior types required by the fea- ture, time_range defines the historical time window considered by the feature, attr_name denotes the specific attributes needed from such behaviors, and comp_func decides the computation function used to summarize behavior attributes. Graph Formulation. The above definition allows us to divide each feature extraction process into four atomic operations, each corresponding to a specific condition to progressively extract neces- sary data from raw app logs. For better understanding, we provide a concrete example in Figure 2 and a high-level illustration in Figure 8. The four atomic operations are as follows: •Retrieve(event_names, time_range): Relevant behavior events re- quired by the feature are retrieved from app logs to device memory according to the specified conditions of <event_names, time_range>. This operation is typically implemented as a database query 2 and data I/O between storage and memory is the primary overhead. •Decode(): For each retrieved row of behavior event, its detailed behavior-specific attributes are decoded from a compressed format. The decoding function is determined by the compression approach during behavior logging, typically implemented with lightweight data transformation tools like JSON parsing [56]. CPU dominates the overhead of this step. •Filter(attr_names): Next, necessary attributes are filtered from the decoded attributes and further converted into a computable format like C array or Python list. •Compute(comp_func): Finally, the filtered attributes are computed into the final input feature using the specified computing function. Common functions include count, average, concatenation to sum- marize user behaviors over a time period in different granularity. Redundancy Identification. In this way, the extraction work- flows of multiple input features can be formulated as one directed acyclic graph. The source node corresponds to raw app log, each target node represents a user feature and they are sequentially con- nected by four critical operation nodes with distinct conditions. Given an FE-graph, redundancy across any features can be identi- fied by directly computing the set intersections of their conditions 2 “SELECT * FROM applog_file WHERE event_name IN event_names AND timestamp > (current_time - time_range)” SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. Event 퐀 User Features Event 퐀 Event 퐀 Conditions of Fused Retrieve Node Outputs of Fused Retrieve Node Divided Retrieve Outputs Decode Outputs Filter Outputs Retrieve Early Branch Decode DecodeFilter Filter Divided Filter Outputs ① Unnecessary Operations ② Repetitive Operations ③ Increased Filtering Costs T i m e _ R a n g e Late Branch Comp ute Feature 1 Feature 2 Feature Union Extra Overhead Comp ute Outputs of Fused Decode Node Outputs of Fused Filter Node Figure 9: Additional cost introduced by feature fusion. for each type of operation nodes. We classify inter-feature redun- dancy into three levels based on the condition overlapping degree. (i) No redundancy: features have completely disjoint <event_names, time_range> conditions, meaning no overlap in their relevant raw data (i.e., non-overlapping event rows in app log); (i) Partial re- dundancy: features share intersected <event_names, time_range> conditions, leading to redundancy in Retrieve and Decode opera- tions; (i) Full redundancy: features have identical <event_names, time_range> conditions, implying duplicate Retrieve and Decode operation costs. It is important to note that few features share to- tally same <event_names, time_range, attr_name> conditions as they depict identical behavior dimensions. 3.3 Graph Optimizer: Inter-Feature Redundancy Elimination After quantifying redundancy across operations, a natural solution to eliminate such redundancy is fusing operation nodes with over- lapping inputs for different features. This involves merging their nodes of the same operation type into a single node, and setting condition of such fused node as the union of the original condi- tions. Despite conceptually simple, this method faces two primary challenges in how to start and terminate the chain fusion process to fully unleash the optimization potential. Overgeneralized Conditions: Chain fusion begins with fusing Re- trieve nodes, which have two orthogonal conditions: <event_names> and <time_range>. Thus, the set union of different Retrieve nodes’ conditions often results in a broader condition scope than the orig- inally intended one. As illustrated in the left part of Figure 9, the condition union of feature 1 and 2 (gray area) is larger than the orig- inally intended condition scope (green and blue areas), which incurs extra operations①on irrelevant data (red area) for all subsequent operation nodes. Optimal Termination Point: Each fused operation chain has to ter- minate with an extra Branch node to separate outputs for different features before their Compute operations, ensuring accurate feature extraction. However, determining the optimal termination point is non-trivial. Early termination could leave redundancy in later operation nodes, while late termination causes excessive branching costs. As shown in Figure 9, early termination after Retrieve node leaves repetitive operations②on overlapping data, whereas delay- ing termination requires each feature to filter relevant data from more intermediate results ③. To address these challenges, AutoFeature introduces a structured optimization strategy, which consists of two key steps: intra-feature 020406080100 Number of Event Entries 10 3 10 2 10 1 10 0 Latency (ms) #Attributes=20 RetrieveDecode 0100200300400500 Number of Event Entries 10 3 10 2 10 1 10 0 10 1 Latency (ms) #Attributes=130 FilterCompute Figure 10: Latency breakdown of extracting user features from behavior events with different attributes. Natural Chronological Order Features with Varied Time Ranges Input Elements of Filter Node Comparison between Each Input and User Feature 1 2345 6 10min 1day10min 1hour1hour1day 1 10min1 hour1 day 3 4 5 2 6 Comparison between Each Input and Lastly Matched Time Range Grouped Time Ranges & Features Figure 11: Comparison between original design and hierar- chical filtering. Each line denotes a comparison. chain partition and inter-feature chain fusion, to eliminate redun- dancy while maximizing computational efficiency. Intra-Feature Chain Partition. The root cause of unnecessary operations during chain fusion lies in the orthogonality of condi- tions of Retrieve nodes (i.e., event_names and time_range). Therefore, AutoFeature proposes to decompose each Retrieve node into multi- ple sub-nodes, where each sub-node retains the original time_range condition but is assigned a distinct event_name condition. This de- composition partitions each feature extraction chain into multiple sub-chains with finer-grained conditions per node, allowing more precise chain fusion. By ensuring that only sub-chains with iden- tical event_name are fused, AutoFeature prevents irrelevant data from entering the pipeline. Inter-Feature Chain Fusion. Further, we introduce “branch postposition” concept to maximize redundancy elimination and a hierarchical filtering algorithm to minimize termination overhead. •Branch Postposition Concept. A key observation from our system evaluation is that Retrieve and Decode nodes dominate feature ex- traction time. As shown in Figure 10, these nodes consume 15× more time than Filter nodes and 300×more time than Compute nodes. To fully eliminating the redundancy of computationally ex- pensive nodes, AutoFeature delays the insertion of Branch nodes until just before the Compute node. •Hierarchical Filtering Algorithm. Instead of appending separate Branch nodes for each feature, AutoFeature proposes to integrate them into the fused Filter node. As a result, for each input element, the Filter node checks whether it satisfies the condition of each fused feature and extracts necessary attributes for different fused features. However, such direct integration increases the Filter node’s computational complexity to푂 푙푒푛(푖푛푝푢푡)×푛푢푚(푓푒푎푡푢푟푒) , which is computationally expensive when facing numerous features or massive relevant behavior events as shown in Figure 11. To reduce this overhead, AutoFeature employs a hierarchical filtering algorithm based on two key observations: (i) Behavior events are stored in chronological order in app log, implicating that the outputs of each operation node are also time-ordered; (i) Model features typically consider meaningful, periodic time ranges Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France (e.g., the past 1 hour, 1 day, 1 week), leading to grouped time_range conditions. Using these properties, AutoFeature could pre-compute a reverse mapping from each time range to the corresponding fea- tures and their required attributes, which can be pre-constructed offline. At runtime, the fused Filter node hierarchically filters at- tributes from input elements for each feature through the following steps, as visualized in Figure 11: •Identify the closest matching time range for each input element based on its timestamp attribute and pre-computed time ranges (i.e., keys of reverse mapping). • Extract necessary attributes for each feature (i.e., values of reverse mapping) associated with the matched time range. Since input elements arrives in chronological order, AutoFeature can begin comparison from the matching time range of previous input element, reducing the overall complexity to푂 푙푒푛(푖푛푝푢푡푠)+ 푙푒푛(푠푒푡(푡푖푚푒_푟푎푛푔푒푠) . This achieves a speedup proportional to the number of fused features even when their time ranges differ. 3.4 Event Evaluator: Inter-Inference Redundancy Minimization In industrial deployment, mobile services require frequent on-device model executions to keep up with the latest user intention and ensure high-quality service. Successive model executions often in- volve extracting user features from overlapping behavior events. To eliminate such temporal redundancy, the classic solution is to cache the valuable intermediate results during feature extraction for future reuse, (i.e., attributes decoded and extracted by each fea- ture. However, this solution faces challenges in real-world mobile apps). Mobile devices typically support multi-app execution, and the memory allocated to one ML model of a single mobile app can be limited and dynamic, making caching all intermediate results not always feasible. To find the sweet spot between memory footprint and redun- dancy elimination, AutoFeature formulates the determination of which intermediate results to cache as a classic knapsack problem. It further incorporates a greedy cache policy to provide theoretically guaranteed performance under various memory budgets. Caching Content Valuation. We first discuss which type of intermediate results should be cached. As analyzed previously, De- code and Retrieve operations dominate the feature extraction cost, which suggests that caching should prioritize these operation nodes’ outputs for maximal computation savings. Therefore, AutoFeature caches at behavior level, i.e., selecting certain behavior types and caching all their events’ necessary attributes, rather than at feature level, i.e., caching attributes for certain features. This approach eliminates the need to re-execute Decode and Retrieve nodes on the same behavior events to extract those uncached attributes. Next, we define two metrics for each behavior type퐸 푖 to quantify its caching utility and cost for further problem formulation: •Utility푈(퐸 푖 )is quantified by the computational savings achieved by caching all necessary attributes of the retrieved events of퐸 푖 , primarily coming from redundant operations on overlapping events between consecutive inferences: 푈(퐸 푖 )= 푁푢푚_푂푣푒푟푙푎푝(퐸 푖 )×퐶표푠푡 _푂푝푡(퐸 푖 ), where푁푢푚_푂푣푒푟푙푎푝denotes the number of overlapped events be- tween consecutive executions and퐶표푠푡 _푂푝푡denotes the operation cost per event. •Cost퐶(퐸 푖 )is the memory space required to cache attributes of events that are processed in current model execution: 퐶(퐸 푖 )= 푁푢푚(퐸 푖 )× 푆푖푧푒(퐸 푖 ), where푁푢푚(퐸 푖 )denotes the number of events processed in current execution and 푆푖푧푒(퐸 푖 ) denotes the size per event. Problem Formulation. Given a set of behavior types퐸 푖 푁 푖=1 and a memory budget푀, we aim to derive the optimal cache pol- icyP ∗ to maximize redundancy reduction while staying within memory constraints: P ∗ =arg max P 푖 ∈0,1 푁 ∑︁ 푖=1 P 푖 ×푈(퐸 푖 ) s.t., 푁 ∑︁ 푖=1 P 푖 ×퐶(퐸 푖 ) ≤푀,(1) whereP 푖 ∈0,1is a decision variable indicating whether to cache the behavior type퐸 푖 . Essentially, the optimization problem (1) is a classic knapsack problem [48,60] that can be solved by dynamic programming algorithm with푂(푁푀)complexity [57]. However, both of the memory constraint푀and the number of overlapped events between model executions are dynamic. They require solv- ing the problem for each real-time model execution and is thus impractical for industrial deployment. Greedy Policy. AutoFeature incorporates a greedy cache policy that prioritizes behavior types based on their utility-to-cost ratio. Specifically, during each feature extraction process, AutoFeature sorts different behavior types by푈(퐸 푖 )/퐶(퐸 푖 )in a descending or- der, and iteratively caches attributes for behavior events with the highest ratio until the memory budget is exhausted. The above greedy policy provides both performance and efficiency guarantees. Theoretically, previous research has proved that greedy solutions can achieve a 2-approximation ratio for the knapsack packing prob- lems [10], which ensures a robust performance across various mem- ory limitations imposed to each ML model. Further, the dynamic ratio푈(퐸 푖 )/퐶(퐸 푖 )can be computed with constant complexity in practice through term decomposition: 푈(퐸 푖 ) 퐶(퐸 푖 ) (푎) = 푇푖푚푒_푂푣푒푟푙푎푝(퐸 푖 )× 퐹푟푒푞(퐸 푖 )×퐶표푠푡 _푂푝푡(퐸 푖 ) 푇푖푚푒_푅푎푛푔푒(퐸 푖 )× 퐹푟푒푞(퐸 푖 )× 푆푖푧푒(퐸 푖 ) = 푇푖푚푒_푂푣푒푟푙푎푝(퐸 푖 ) 푇푖푚푒_푅푎푛푔푒(퐸 푖 ) | z Dynamic Term 1 × 퐶표푠푡 _푂푝푡(퐸 푖 ) 푆푖푧푒(퐸 푖 ) | z Static Term 2 , where Equation (a) represents the number of events as the mul- tiplication of time range and the behavior occurrence frequency. As a result, term 1 is dynamically determined by on-device model inference frequency and term 2 is static and can be recorded once in an offline manner. Online Execution. To integrate this caching policy into on- line feature extraction, AutoFeature deploys an event evaluator that dynamically adjusts caching decisions at runtime. The fea- ture extraction workflow follows these steps: (i) Retrieve attributes of relevant behavior events from cache and update the feature’s time_range conditions based on the timestamp of cached events; (i) Execute Retrieve and Decode operation nodes for missing attributes of newly logged events, (i) Merge cached and newly extracted SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. attributes and execute Filter and Compute nodes to generate final features, (iv) Greedily cache events based on their up-to-date ratios of utility and cost, which are also leveraged to update cache when facing dynamic memory budget. 4 Evaluation 4.1 Experiment Setup Implementation. We have implemented a system prototype of AutoFeature and integrated it into the SDK of industrial mobile apps for evaluation 3 . AutoFeature is designed to operate without modifying the model inference process, ensuring that model train- ers, inference engine developers and mobile users do not need to make any adjustments. In compliance with enterprise data privacy requirements, our evaluation primarily uses ByteNN [54] as back- end engine, which delivers comparable or superior performance to the open-source MNN [33] through app-specific optimizations. Mobile Services. Our evaluation was conducted on popular mobile apps from ByteDance, a company with billions of users, advanced on-device AI infrastructure and diverse mobile services. To demonstrate the generality of AutoFeature, we evaluate its per- formance across five representative mobile services, spanning three typical domains of mobile apps: search engines, video apps and e-commerce platforms. For video domain, TikTok (Douyin) is one of the world’s most popular short-form video platforms [71]. For search domain, Toutiao [72] is one of the most popular content discovery and searching platforms in China. For, e-commerce do- main, Douyin is also a popular e-commerce platform. The tested services vary widely in scenarios, tasks, input user features and model execution frequencies, which are visualized in Figure 12. •Content Preloading (CP): Common in video apps like TikTok, this service decides which segment of video and comment to preload for ensuring seamless video watching experiences. The model relies on 86 user features derived from 27 distinct behavior types (e.g., video click, playback duration, sharing). •Keyword Prediction (KP): Deployed in mobile search engines like Google, this service predicts likely search query keywords based on past query behavior and currently browsing content. It uses 53 user features to track 22 types of search-related behaviors (e.g., past search terms, click-through history, time spent on search results.). •Search Ranking (SR): This service ranks the returned search results to match current user preferences and improve searching relevance. It uses 40 user features to tracks 10 user behavior types related to searches such as interactions with ranked elements and engagement with search results. •Product Recommendation (PR): Widely used in e-commerce plat- forms like Taobao, this service offers personalized product recom- mendations and targeted advertisements. Its model extracts 103 user features across 21 types of commercial behaviors, including product browsing duration, cart additions, purchases, and search within product categories. •Video Recommendation (VR): generates personalized video sug- gestions based on a user’s viewing history and content preferences. 3 Updates to the model configurations like input features and model parameters are typically handled by downloading a new mobile app SDK through standard app updates. For this case, AutoFeature simply treats the updated configuration as a new model and re-runs its offline optimization phase within millisecond-level latency. The relevant user features and behavior types for this model were previously illustrated in Figure 6(a). As (i) modern mobile operating systems aggressively prioritize the currently active foreground app and (i) on-device model inferences are mostly triggered when the user is interacting with the app, our testing app is guaranteed to be in the foreground and receive the dominant share of available hardware resources, limiting the interference from other co-running background apps. Model Architecture. While we cannot disclose specific model architectures due to confidential requirement, we present a general structure of on-device models used by popular mobile services in Figure 13, which composes of three layers: •Input Layer: As analyzed in §2.1, An on-device model takes three categories of features as inputs: cloud features to provide global information, device features to describe the current device state and user features to summarize various historical user behaviors. •Processing Layer: These features are then processed by different model layers. Statistical features of user behaviors and device fea- tures are passed into an factorization machine layer for feature crossing, while sequential behavior features are sent to a sequence encoder to capture temporal dynamics and periodical patterns. •Output Layer: Finally, the combined feature outputs are passed through several dense layers to generate final predictions for per- sonalized system responses. Testing Users. We evaluate AutoFeature’s performance through online evaluation within real-world mobile apps. For each mobile service, we collect the end-to-end latency of on-device model exe- cution of 10 testing users during their daily usage of the app across 2 days, containing three common time periods: noon (12:00-13:00), evening (18:00-19:00) and night (21:00-23:00). The limited test group stems from a necessary trade-off between operational cost and data representativeness. A fair and reliable comparison requires repeat- edly running different feature extraction methods for every single real-time inference request. While this ensures identical user data and device state across methods, it significantly degrades service responsiveness and user experience. Limiting testing to 10 anony- mous real-world users mitigates operational cost and financial loss, while still ensuring representativeness of the general population: •Statistically Similar Usage Patterns: We quantitatively validate that the test group’s statistical usage patterns closely match those of massive real-world active users. Figure 14 visualizes the behav- ior frequencies of thousands of real-world user base (black bars) against 10 test users’ 20 traces (red stars). We used the Kolmogorov- Smirnov (KS) test to compare these distributions. For different time periods, the KS statistic is extremely low (0.079-0.118) and the p- values (0.785-0.998) are significantly greater than the standard 0.05 threshold, confirming that there is no statistically significant differ- ence between the usage patterns of test users and overall user base. •Diverse Behavior Traces: Beyond statistical alignment, our 10 test users were selected to cover a wide range of activity intensity, demonstrating the generality of our evaluation. Figure 15 illustrates the fine-grained behavior frequencies of test users, segmented into 10-minute intervals across different time periods.The top 10% most active users (P90 traces) generate over 45 specific behaviors every 10 minutes, while the bottom 30% of users (P30 traces) generate fewer than 5 behaviors per 10 minutes. Further details are provided in Appendix A. Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France 01020304050607080 1min 1hour 1day 1week confirm_A play_A poi_A life_A operation_A life_Blife_C life_D favourite_A project_A project_B enter_A anchor_A video_A poi_B click_A order_A Content Preloading 01020304050 1min 1hour 1day 1week search_A PRE_A trending_A choose_A click_A click_B search_B leaderboard_A click_C favourite_A poi_A poi_B search_C enter_A enter_B follow_A project_A post_A share_A trending_B search_D play_A Keyword Prediction 0510152025303540 ID of User Feature 1min 1hour 1day 1week video_A like_A play_A share_A speed_A favourite_A Search Ranking 020406080100 ID of User Feature 1min 1hour 1day 1week project_A live_A livesdk_A live_B click_A enter_A click_B mplive_A project_B poi_A favourite_A live_C livesdk_B click_C live_D click_Denter_B poi_B play_A homepage_A Time Range Product Recommendation (a) User feature statistics: Each bar denotes a feature and each color represents a behavior type. 2s10s1min10min Inference Interval 0.00 0.25 0.50 0.75 1.00 CDF across Inferences KP CP SR PR VR (b) Inference frequency distribution. Figure 12: User feature definitions and online execution frequencies of on-device ML models in our evaluation. User&Service Embedding Context Features ... Cloud Features Battery Level Network Status ... Device Features Long-Term Statistics User Behavior Features Short-Term Statistics Sequence Encoder FM Feature Cross Layer Behavior Sequences Dense Layer Whether to Pre-load Content Model Outputs for System Response ... Sorted Products for RecommendationPredicted Searching Keyword Dense Layer Figure 13: Model structure of common on-device ML models. 05001000150020002500 0 100 200 300 400 500 Behavior Frequencies KS Statistic: 0.0788 P-value: 0.9894 Noon: 12:00-13:00 Tesing Users Online Users 05001000150020002500 0 200 400 600 800 KS Statistic: 0.0693 P-value: 0.9977 Evening: 18:00-19:00 Tesing Users Online Users 050010001500200025003000 0 200 400 600 800 KS Statistic: 0.1175 P-value: 0.7850 Trace ID of Online Users (or Aligned ID of Testing Users) Night: 21:00-23:00 Tesing Users Online Users Figure 14: Statistically similar usage patterns between testing users and thousands of real-world users. 1213 1819 21 2223 Time Segments across 3 Periods 0 15 30 45 60 Behavior Num. (per 10 min) P90 P80 P70 P60 P50 P30 Mean Figure 15: Specific behavior traces of testing users. It is important to note that all data collection and measure- ment processes are conducted with the consent of testing users, ensuring full compliance with privacy regulations in both research and industry. Baselines. To the best of our knowledge, AutoFeature is the first work to optimize feature extraction for on-device model exe- cution of real-world mobile apps. Thus, we consider the following baselines and ablated versions of AutoFeature: (i) w/o AutoFeature: the industry-standard on-device feature extraction process, where each user feature is extracted independently without optimiza- tion, and (i) w/ Fusion: only the graph optimizer of AutoFeature is employed to eliminate inter-feature redundant operations, (i) w/ Cache: only the cache policy of AutoFeature is applied to minimize inter-inference redundancy. 4.2 Overall Performance We begin by evaluating the overall performance of AutoFeature across diverse mobile services. Figure 16 plots the model execution latencies of different on-device ML models in varied time periods, which are averaged on testing users. AutoFeature significantly accelerates end-to-end on-device model execution. For mobile services in various domains, Aut- oFeature consistently reduces the on-device model execution la- tency to below 20ms, which exceeds the human perception range of 30-60 FPS and successfully becomes imperceptible to users. Com- pared to the original industry-standard feature extraction process (w/o AutoFeature), AutoFeature achieves inference speedups rang- ing from 1.72-3.44×for CP service, 1.33-1.44×for KP service, 1.41- 4.53×for SR service, 1.82-2.18×for PR service and 3.93-4.43×for VR service. Additionally, we observe that AutoFeature’s performance is closely related to the original execution latency (i.e., w/o AutoFea- ture): the largest speedups are observed in CP, SR and VR models with average latency exceeding 40 ms, while the lowest speedup is achieved by KP model with around 20 ms latency. The reason is that AutoFeature reduces the practical feature extraction latency to relatively low and leaves model inference time as the dominant factor that limits the end-to-end model execution acceleration. The performance of AutoFeature continually improves with time. Interestingly, from Figure 16, we notice that AutoFea- ture delivers higher speedups of model execution at night compared to daytime periods, which holds across nearly all testing mobile ser- vices. For example, for CP service, AutoFeature achieves speedups of 1.72×and 2.79×during noon and evening, but reaches as high as 3.44×at night. Upon deeper analysis, we attribute this perfor- mance variation to common patterns of user behaviors throughout a day. At night, users tend to engage more actively with mobile apps over an extended and uninterrupted period. This results in a higher volume of newly logged behavior events, increasing the opportunity for AutoFeature to optimize feature extraction by elim- inating redundant computations. In contrast, during midday and evening breaks, user interactions with mobile apps are typically shorter and less frequent, due to normal work schedules. The contributions of redundancy elimination across fea- tures and inferences vary across mobile services. Figure 16 also reveals that both the graph optimizer (w/ Fusion) and cache policy (w/ Cache) could accelerate on-device model execution for all mo- bile services, except for the first model execution during each time period as app exit frees up memory. However, their effectiveness varies significantly across different services. While the w/ Fusion plays a dominant role in accelerating CP, KP, and PR model exe- cutions, it is less effective for SR and VR models. This discrepancy primarily arises from the differences in their overlapping degrees of behavior events across models. Specifically, as shown in Figure SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. 1213 1819 212223 Wall-Clock Time 10 20 30 40 50 Execution Latency (ms) NoonEveningNight 1.72x2.79x3.44x Content Preloading (CP) 1213 1819 212223 Wall-Clock Time 10 15 20 25 NoonEveningNight 1.33x1.42x1.44x Keyword Prediction (KP) 1213 1819 212223 Wall-Clock Time 5 20 35 50 65 NoonEveningNight 1.41x3.00x4.43x Search Ranking (SR) 1213 1819 212223 Wall-Clock Time 10 15 20 25 30 35 NoonEveningNight 1.82x2.03x2.18x Product Recommendation (PR) w/o AutoFeaturew/o Cachew/o Fusionw/ AutoFeature 1213 1819 212223 Wall-Clock Time 5 20 35 50 65 NoonEveningNight 3.93x4.49x4.27x Video Recommendation (VR) (a) End-to-end latency of on-device model execution of different methods. We annotate the average speedups achieved by AutoFeature. 1213 1819 212223 Wall-Clock Time 1 2 3 4 Execution Speedup (x) Content Preloading (CP) 1213 1819 212223 Wall-Clock Time 1.0 1.2 1.4 1.6 1.8 Keyword Prediction (KP) 1213 1819 212223 Wall-Clock Time 1 2 3 4 5 Search Ranking (SR) 1213 1819 212223 Wall-Clock Time 1.0 1.5 2.0 2.5 Product Recommendation (PR) w/o AutoFeaturew/o Cachew/o Fusionw/ AutoFeature 1213 1819 212223 Wall-Clock Time 1 2 3 4 5 Video Recommendation (VR) (b) Speedups of on-device model execution of different methods compared with industry-standard baseline (w/o AutoFeature). Figure 16: Overall performance of different methods across varied time ranges and mobile services. KPCPSRPRVR Mobile Services 0 1 2 3 4 Offline Latency (ms) 0.19 3.32 0.21 2.71 0.09 1.23 0.26 3.15 0.20 3.01 FE-Graph Opt.Event Profile (a) Offline: time consumption. KPCPSRPRVR Mobile Services 0 25 50 75 100 Extra Memory (KB) 15.9 39.6 57.5 23.3 85.8 (b) Online: memory footprint. Figure 17: System over- heads introduced by AutoFeature. 12(a), 80.2% of features in CP, 85% in KP, and 80.6% in PR share iden- tical event_name conditions, i.e., the same relevant behavior types, which enables AutoFeature to eliminate a substantial portion of redundant computations across features. In contrast, the SR and VR models exhibit relatively lower overlap (59% and 71%, respectively), reducing the optimization potential of inter-feature fusion. AutoFeature introduces marginal extra system costs for both offline and online phases. During the offline optimization phase, the primary system overhead comes from (i) constructing and optimizing the FE-graph for each model, and (i) profiling the operation cost and result size for each interaction event type. As shown in Figure 17(a), the overall time cost of offline optimization is dominated by the profiling process, which ranges from 1.23 ms to 3.32 ms across different on-device ML models. Note that the memory consumption during this phase is negligible, as AutoFea- ture requires only a few kilobytes space to analyze the conditions across features. During the online execution phase, the extra sys- tem overhead stems from caching intermediate results to reduce redundant operations across consecutive inferences. As depicted in Figure 17(b), the average extra memory footprint required to cache all intermediate results remains consistently below 100 KB. Such low memory cost is generally acceptable for both high-end and low-end mobile devices, attributed to (i) the compact feature sets of lightweight on-device ML models, (i) our efficient event- level data caching for all features. As a result, our proposed greedy cache policy is mainly utilized when the memory budget of each ML model is strictly limited by operating system, analyzed in §4.3. Comparison with Cloud-based Methods We further bench- mark AutoFeature against two cloud-side feature extraction sys- tems. They trade real-time computation for increased storage by offloading expensive operations to an offline logging process and maintaining an additional database to store pre-computed outputs. Since퐷푒푐표푑푒and푅푒푡푟푖푒푣푒are the top-2 computationally expensive operations, we implemented two baselines based on which opera- tion to offload: Decoded Log and Feature Store. Table 1 elaborates their storage structures and offloaded operations. Figure 18(a) plots the average inference latency across various mobile services. AutoFeature significantly reduces inference la- tency across the five services by up to 70.9%, 30.5%, 77.4%, 54.2% and 77.7%, and reduces latency by up to 31.35 ms, 5.85 ms, 38.76 ms, 15.68 ms and 38.94 ms. Offloading only퐷푒푐표푑푒(Decoded Log) yields an additional gain of at most 4.38 ms, and offloading both 퐷푒푐표푑푒and푅푒푡푟푖푒푣푒(Feature Store) further reduces latency by at most 3.91 ms. However, the cloud baselines accelerate feature extraction with a catastrophic increase in storage requirements, making them impractical for mobile deployment. Figure 18(b) illus- trates the app log size distributions across testing users for each system. For an average user, Decoded Log increases the app log size by 2.61×, and Feature Store increases it by a staggering 2.80×. This massive storage inflation is unacceptable for production mobile apps due to its direct link to user churn and financial loss: (i) Public statistics confirm that excessive app size is a primary driver of app uninstallation [3]; (i) Our internal industrial data reveals that every additional 10 MB in app size leads to a decrease of around 30,000 to 61,000 daily active users, resulting in a daily financial loss of over $7,000. The storage efficiency provided by AutoFeature is thus a prerequisite for real-world deployment. 4.3 Component-Wise Analysis Next, we evaluate the effectiveness of each key design in AutoFea- ture. The experiments are mainly conducted on the video recom- mendation service with the most complex feature dependencies. Inter-Feature Fusion. To assess the impact of inter-feature operation fusion, we conduct a breakdown analysis of the feature extraction latency, isolating the time cost of each operation. Figure 19(a) compares the latency distributions across online executions before and after fusing redundant operations. Our key finding is that the primary computational savings of inter-feature fusion stem from two bottleneck operations: Decode and Retrieve. Specifically, the average latency of Decode decreases from 12.01 ms to 2.95 ms and Retrieve decreases from 9.12 ms to 2.23 ms, yielding over 4× speedups for both of them. However, the fusion process slightly Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France Table 1: Detailed introduction to cloud-side feature extraction baselines. System Offloaded Operations Storage StructureIntroduced Storage AutoFeature✘ Each Row: One Behavior Event ✘ Each Column: Compressed Attributes Decoded LogDecode Each Row: One Behavior Event Massive Columns Each Column: One Unique Attribute Feature Store DecodeEach Row: One Behavior Event required by One FeatureRedundant Rows RetrieveEach Column: One Unique AttributeMassive Columns Noon EveningNight 0 10 20 30 40 50 Inference Latency (ms) 21.39 9.52 8.22 12.48 -41.7% 35.26 11.93 9.36 12.53 -64.2% 44.24 13.49 10.11 12.79 -70.9% Content Preloading w/o AutoFeatureDecoded LogFeature Storew/ AutoFeature Noon EveningNight 0 5 10 15 20 25 17.45 8.84 7.89 13.24 -24.7% 18.80 9.07 8.00 13.07 -29.8% 18.98 9.10 8.02 13.27 -30.5% Keyword Prediction Noon EveningNight 0 10 20 30 40 50 15.30 8.47 7.71 10.94 -28.9% 33.65 11.65 9.23 10.97 -66.6% 50.11 14.51 10.59 11.03 -77.4% Search Ranking Noon EveningNight 0 6 12 18 24 30 23.73 9.93 8.41 12.94 -45.1% 26.49 10.41 8.64 13.02 -50.7% 28.90 10.82 8.84 13.15 -54.2% Product Recommendation Noon EveningNight 0 12 24 36 48 60 39.12 12.60 9.68 9.73 -74.5% 49.34 14.37 10.53 10.55 -77.7% 46.27 13.84 10.27 10.34 -76.6% Video Recommendation (a) Inference Latency. 01020304050 Log Size (MB) 0.00 0.25 0.50 0.75 1.00 CDF across Users w/ AutoFeature Decoded Log Feature Store (b) Storage Cost. Figure 18: Performance comparison between AutoFeature and cloud-side feature extraction baselines. RetrieveDecodeFilter 0.01 0.1 1.0 10 Time Cost (ms) 12.05 9.10 0.05 2.95 2.20 0.07 Before After (a) Effect of Inter-Feature Fusion 0.000.250.500.751.00 Relative Memory Budget 0.00 0.25 0.50 0.75 1.00 Inter-Execution Redund. w/ Random AutoFeature (b) Effect of Cache Policy. Figure 19: Component-wise analysis. 10s1m10m0.5h Inference Interval 1.5 2.0 2.5 3.0 3.5 4.0 Execution Speedup ( × ) CP KP SR PR VR Figure 20: Impact of inference interval across services. 0.00.20.40.60.8 Inter-Feature Redundancy 1 100 200 300 AutoFeature Speedup ( × ) High-Frequency Case 10s 1min 5min 0.00.20.40.60.8 1 20 40 60 Low-Frequency Case 30min 1hour 1day Figure 21: Impact of inter-feature redun- dancy through offline evaluation. increases the latency of Filter operation. This is in expectation because the fused Filter operation has to process and differentiate outputs for each fused feature, introducing extra branching cost. Benefited from our proposed hierarchical filtering algorithm, the extra latency is reduced to only 0.02 ms on average. Inter-Execution Optimization. Next, we evaluate the per- formance of greedy cache policy in reducing redundancy across model executions with various memory constraints. We compare AutoFeature with a modified version (w/ Random), which caches in- termediate results of different behavior types randomly rather than utility-to-cost ratio. Because (i) the redundancy between consecu- tive inferences varies during online evaluation and (i) the number of intermediate results differs across mobile users, we analyze the relative changes in redundancy as a function of the proportion of intermediate results cached by the device. As shown in Figure 19(b), greedy cache policy consistently outperforms the w/ Random, demonstrating its robust performance in handling various memory budgets. We also notice that the greedy policy is more effective when device memory is limited, such as reducing 50% redundant feature extraction operations by caching only 23% intermediate results. This is because the behavior types prioritized by AutoFea- ture exhibit higher utility-to-cost ratio, implying more redundancy reduction per memory cost unit, highlighting AutoFeature’ superior performance when handling strictly limited device memory. 4.4 Sensitivity Analysis Since AutoFeature operates without hyper-parameters, eliminating the need for trial-and-error tuning, we focus our sensitivity analysis on two environmental factors. Impact of Model Execution Frequency. To evaluate how execution frequency affects AutoFeature’s performance, we system- atically vary the intervals between consecutive model executions. Specifically, we force on-device ML models of different mobile ser- vices to be invoked at fixed intervals during a user’s app usage at night. As shown in Figure 20, the speedups achieved by AutoFea- ture decrease as the interval between model executions increases. This trend is expected, as longer intervals reduce the likelihood of overlapping behavior events processed by consecutive feature ex- tractions, thereby limiting the potential for redundancy elimination. However, even under infrequent scenarios (e.g., 1 model execution per 30 minutes), AutoFeature still achieves significant speedups ranging from 1.40×to 2.8×across real-world mobile services. This result demonstrates AutoFeature’s robustness in handling diverse inference frequencies. Impact of Feature Redundancy. To further assess AutoFea- ture’s performance across potential ML models with varying levels of inter-feature redundancy, we construct numerous synthetic fea- ture sets with controlled redundancy levels. Specifically, we define feature redundancy as the proportion of overlapping time ranges among features that rely on the same user behavior types. For each redundancy level, we measure AutoFeature’s speedups on model execution under two scenarios: (i) high-frequency inferences with intervals ranging from 10 seconds to 5 minutes, and (i) low- frequency inferences with intervals ranging from 30 minutes to 1 day. As shown in Figure 21, AutoFeature’s performance consis- tently improves with increasing feature redundancy across various inference frequencies. When inference requests occur every 10s and 1h, the feature extraction speedups grow from 7.3×and 1.0×at SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. 0% redundancy to 336×and 21.9×at nearly 90% redundancy. Even for the longest inference interval of 1 day, AutoFeature still reduces feature extraction latency by 2.1×, 4.1×and 5.6×for redundancy levels of 20%, 50% and 80%. Te observed speedups do not increase linearly with redundancy levels, as the plotted feature redundancy is estimated based on overlapping time ranges rather than specific behavior events, which are dynamic with user behaviors. It is im- portant to note that the speedups reported in Figure 21 exceed those observed in online evaluations, as the synthetic experiments isolate and measure only the feature extraction process, whereas online evaluations include the model inference latencies. 5 Discussion on Limitations Model-Engine Co-Design. In practice, the ML development pipeline is often divided between two distinct teams: the Model Team, which focuses on maximizing model accuracy and company profit, and the Infrastructure Team, which focuses on optimizing model ex- ecution speed. While this decoupling makes our team’s feature extraction engine model-agnostic and highly generalizable, it in- herently limits the opportunities for model-engine co-design. A co-designed approach could unlock greater efficiency by trading a small, acceptable amount of accuracy for significant latency gains, such as reusing stale feature values rather than recomputing the fresh ones. However, these optimizations are currently hindered by the prevailing organizational silos and technical separation typical of mobile application companies. Dependency on Active Users and App Services. The appli- cability of AutoFeature is naturally limited to user scenarios and application types where feature extraction constitutes a measur- able performance bottleneck. For inactive users with minimal app logs, the feature extraction overhead might be minimal and lower than model inference time, as only a few behaviors are processed. However, this limitation applies only to an edge case, as nearly all major mobile services prioritize optimizing the experience of active users, who are the primary drivers of company revenue and inference requests. Also, for future intensive services where models become larger, the model inference time can dominate the exe- cution pipeline. While this will diminish the proportional impact of our optimization on the total latency, we believe that the need for real-time responsiveness will remain paramount for the main mobile services and make AutoFeature still applicable. 6 Related Work Device-Side Inference Acceleration. Extensive research has been conducted to optimize on-device model inference, which can be classified into four categories: operator optimization [33, 38,51] to enhance backend inference engines at the operator level, model architecture optimization through quantization [37,42,45], pruning [47,52,63,70], sparsification [8,46] and architecture re- design [12,28,65] to reduce computational complexity, hardware resource exploitation [9,31,32,36,69,74] to utilize multiple types of on-device computational hardware to accelerate computation, and inference frequency reduction [11,29,41,77] to avoid unnecessary inference requests. However, these work primarily focused on opti- mizing the model execution stage due to exclusively considering traditional vision or language models that use static input features. As a result, AutoFeature is complementary to them by optimizing the previous feature extraction stage for practical on-device ML adoption in mobile apps. Cloud-Side Feature Computation. Another related research topic is feature computation at cloud servers. To provide compre- hensive features for real-time model inference processes requested by different online services, many service providers maintain an up- to-date feature store at cloud servers like Amazon SageMaker [6] and Databricks Feature Store [15] , which collects all user data, pre-computes and stores necessary user features. This setups al- lows multiple application services to concurrently query user fea- tures [34,73], but relies on huge storage and computational ca- pabilities of cloud servers and introduces significant privacy con- cerns [49,68]. Consequently, there is a growing trend towards offloading both feature extraction and model execution processes to device for more real-time, secure and personalized serves [25,75]. Our work analyzes the practical issues in on-device feature extrac- tion process and further explores potential optimization directions. Feature Selection is a mature field focused on identifying the optimal subset of features to maximize accuracy and minimize dimensionality [40,78]. To achieve this, filter methods valuate fea- tures using a scoring function based on statistical properties like mutual information gain [18, 21, 64, 66] and representativeness to data distribution [19,27,30] . Wrapper methods leverage the model performance to assess the quality of selected features and refin- ing the selection iteratively [5,62,79]. However, feature selection operates before deployment to determine what inputs the model should use, and our work operates in the post-deployment stage to optimize how those essential features are extracted for real-time on-device model inferences 7 Conclusion In this work, we identify an overlooked bottleneck of feature extrac- tion in on-device model inference with user behavior sequences, which are prevalent in real-world industrial mobile apps. To address this bottleneck, we propose AutoFeature, the first feature extraction engine designed to accelerate on-device execution by eliminating redundant operations across both input features and consecutive model executions without compromising model accuracy. We imple- ment a system prototype and integrate it into five real-world mobile services for evaluation, where AutoFeature achieves 1.33×-4.53× speedup in end-to-end model execution latency. Acknowledgments This work was supported in part by National Key R&D Program of China (No. 2023YFB4502400), in part by China NSF grant No. 62322206, 62432007, 62441236, 62132018, 62025204, U2268204, 62272307, 62372296, U25A6024, in part by Fundamental and Interdisciplinary Disciplines Breakthrough Plan of the Ministry of Education of China (No. JYB2025XDXM103). The opinions, findings, conclusions, and recommendations expressed in this paper are those of the authors and do not necessarily reflect the views of the funding agencies or the government. Zhenzhe Zheng is the corresponding author. References [1] [n. d.]. 2024 Q1 Mobile Application Performance Experience Report. https: //lydaas.oss-cn-beijing.aliyuncs.com/doc/2024Q1_U-APM.pdf . Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France [2]2021. On-device training with tensorflow lite. https://w.tensorflow.org/lite/ examples/on_device_training/overview. [3]2025. App uninstall report – 2025 edition. https://w.appsflyer.com/resources/ reports/app-uninstall-benchmarks/. [4]Mário Almeida, Stefanos Laskaridis, Abhinav Mehrotra, Lukasz Dudziak, Ilias Leontiadis, and Nicholas D. Lane. 2021. Smart at what cost?: characterising mobile deep neural networks in the wild. In ACM Internet Measurement Conference (IMC). 658–672. [5] Hiromasa Arai, Crystal Maung, Ke Xu, and Haim Schweitzer. 2016. Unsupervised feature selection by heuristic search with provable bounds on suboptimality. In Proceedings of the AAAI conference on artificial intelligence, Vol. 30. [6]AWS. 2025. Amazon SageMaker Feature Store: A fully managed service for machine learning features. https://aws.amazon.com/sagemaker/ai/feature-store/. [7]Nicholas J Belkin and W Bruce Croft. 1992. Information filtering and information retrieval: Two sides of the same coin? Commun. ACM 35, 12 (1992), 29–38. [8]Sourav Bhattacharya and Nicholas D Lane. 2016. Sparsification and separation of deep learning layers for constrained resource inference on wearables. In ACM Conference on Embedded Networked Sensor Systems (Sensys). 176–189. [9]Qingqing Cao, Niranjan Balasubramanian, and Aruna Balasubramanian. 2017. MobiRNN: Efficient Recurrent Neural Network Execution on Mobile GPU. In Proceedings of the 1st International Workshop on Embedded and Mobile Deep Learning (Deep Learning for Mobile Systems and Applications). 1–6. [10]Chandra Chekuri and Sanjeev Khanna. 2005. A polynomial time approximation scheme for the multiple knapsack problem. SIAM J. Comput. 35, 3 (2005), 713–728. [11]Tiffany Yu-Han Chen, Lenin Ravindranath, Shuo Deng, Paramvir Bahl, and Hari Balakrishnan. 2015. Glimpse: Continuous, real-time object recognition on mobile devices. In ACM Conference on Embedded Networked Sensor Systems (Sensys). 155–168. [12]Yanjiao Chen, Baolin Zheng, Zihan Zhang, Qian Wang, Chao Shen, and Qian Zhang. 2020. Deep learning on mobile and embedded devices: State-of-the-art, challenges, and future directions. ACM Computing Surveys (CSUR) 53, 4 (2020), 1–37. [13]Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, et al. 2016. Wide & deep learning for recommender systems. In Proceedings of the 1st workshop on deep learning for recommender systems. 7–10. [14] Paul Covington, Jay Adams, and Emre Sargin. 2016. Deep neural networks for youtube recommendations. In Proceedings of the 10th ACM conference on recommender systems. 191–198. [15] databricks. 2025. Databricks Feature Store. https://docs.databricks.com/gcp/en/ machine-learning/feature-store. [16] Michael F Deering. 1998. The limits of human vision. In 2nd international immer- sive projection technology workshop, Vol. 2. [17] Apple Developer. 2023. Maximum build file sizes. https://developer.apple.com/ help/app-store-connect/reference/maximum-build-file-sizes/. [18] Liang Du, Zhiyong Shen, Xuan Li, Peng Zhou, and Yi-Dong Shen. 2013. Local and global discriminative learning for unsupervised feature selection. In E International Conference on Data Mining (ICDM). 131–140. [19] Ahmed K Farahat, Ali Ghodsi, and Mohamed S Kamel. 2011. An efficient greedy method for unsupervised feature selection. In IEEE nternational Conference on Data Mining (ICDM). IEEE, 161–170. [20]Jesse Feiler and Jesse Feiler. 2015. Using SQLite with Core Data (iOS and OS X). Introducing SQLite for Mobile Developers (2015), 61–73. [21] Shuyang Gao, Greg Ver Steeg, and Aram Galstyan. 2016. Variational information maximization for feature selection. Advances in neural information processing systems (NeurIPS) 29 (2016). [22]Chen Gong, Zhenzhe Zheng, Yunfeng Shao, Bingshuai Li, Fan Wu, and Guihai Chen. 2024. ODE: An online data selection framework for federated learning with limited storage. IEEE/ACM Transactions on Networking 32, 4 (2024), 2794–2809. [23]Chen Gong, Zhenzhe Zheng, Fan Wu, Xiaofeng Jia, and Guihai Chen. 2024. Delta: A cloud-assisted data enrichment framework for on-device continual learning. In Proceedings of the 30th Annual International Conference on Mobile Computing and Networking (MobiCom). 1408–1423. [24]Chen Gong, Zhenzhe Zheng, Fan Wu, Yunfeng Shao, Bingshuai Li, and Guihai Chen. 2023. To store or not? online data selection for federated learning with limited storage. In Proceedings of the ACM Web Conference 2023 (W). 3044– 3055. [25]Chen Gong, Yan Zhuang, Zhenzhe Zheng, Yiliu Chen, Sheng Wang, Fan Wu, and Guihai Chen. 2025. Optimizing Storage Overhead of User Behavior Log for ML-embedded Mobile Apps. Proceedings of the ACM on Measurement and Analysis of Computing Systems (SIGMETRICS) 9, 3 (2025), 1–28. [26]Google. [n. d.]. Android Developers: APK Expansion Files. https://developer. android.com/google/play/expansion-files. [27]Quanquan Gu, Zhenhui Li, and Jiawei Han. 2012. Generalized fisher score for feature selection. arXiv preprint arXiv:1202.3725 (2012). [28]Peizhen Guo, Bo Hu, and Wenjun Hu. 2021. Mistify: AutomatingDNNModel Porting forOn-DeviceInference at the Edge. In USENIX Symposium on Net- worked Systems Design and Implementation (NSDI). 705–719. [29]Peizhen Guo, Bo Hu, Rui Li, and Wenjun Hu. 2018. Foggycache: Cross-device approximate computation reuse. In International Conference on Mobile Computing and Networking (MobiCom). 19–34. [30] Xiaofei He, Deng Cai, and Partha Niyogi. 2005. Laplacian score for feature selection. Advances in neural information processing systems (NeurIPS) 18 (2005). [31]Joo Seong Jeong, Jingyu Lee, Donghyun Kim, Changmin Jeon, Changjin Jeong, Youngki Lee, and Byung-Gon Chun. 2022. Band: coordinated multi-dnn infer- ence on heterogeneous mobile processors. In International Conference on Mobile Systems, Applications and Services (MobiSys). 235–247. [32]Fucheng Jia, Deyu Zhang, Ting Cao, Shiqi Jiang, Yunxin Liu, Ju Ren, and Yaoxue Zhang. 2022. CoDL: efficient CPU-GPU co-execution for deep learning inference on mobile devices.. In International Conference on Mobile Systems, Applications, and Services (MobiSys), Vol. 22. 209–221. [33]Xiaotang Jiang, Huan Wang, Yiliu Chen, Ziqi Wu, Lichuan Wang, Bin Zou, Yafeng Yang, Zongyang Cui, Yu Cai, Tianhang Yu, et al.2020. MNN: A universal and efficient inference engine. Proceedings of Machine Learning and Systems (MLSys) 2 (2020), 1–13. [34] Theofilos Kakantousis, Antonios Kouzoupis, Fabio Buso, Gautier Berthou, Jim Dowling, and Seif Haridi. 2019. Horizontally scalable ml pipelines with a feature store. In Proceedings of the 2nd SysML Conference. [35]Shubhra Kanti Karmaker Santu, Parikshit Sondhi, and ChengXiang Zhai. 2017. On application of learning to rank for e-commerce search. In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR). 475–484. [36]Mehrdad Khani, Ganesh Ananthanarayanan, Kevin Hsieh, Junchen Jiang, Ravi Netravali, Yuanchao Shu, Mohammad Alizadeh, and Victor Bahl. 2023.RECL: ResponsiveResource-Efficientcontinuous learning for video analytics. In USENIX Symposium on Networked Systems Design and Implementation (NSDI). 917–932. [37] Youngsok Kim, Joonsung Kim, Dongju Chae, Daehyun Kim, and Jangwoo Kim. 2019.휇layer: Low latency on-device inference using cooperative single-layer acceleration and processor-friendly quantization. In Proceedings of the Fourteenth EuroSys Conference (EuroSys). 1–15. [38]Rui Kong, Yuanchun Li, Yizhen Yuan, and Linghe Kong. 2023. Convrelu++: Reference-based lossless acceleration of conv-relu operations on mobile cpu. In International Conference on Mobile Systems, Applications and Services (MobiCom). 503–515. [39] Jay Kreibich. 2010. Using SQLite. " O’Reilly Media, Inc.". [40] Jundong Li, Kewei Cheng, Suhang Wang, Fred Morstatter, Robert P Trevino, Jiliang Tang, and Huan Liu. 2017. Feature selection: A data perspective. ACM computing surveys (CSUR) 50, 6 (2017), 1–45. [41]Yuanqi Li, Arthi Padmanabhan, Pengzhan Zhao, Yufei Wang, Guoqing Harry Xu, and Ravi Netravali. 2020. Reducto: On-camera filtering for resource-efficient real-time video analytics. In Proceedings of the Annual conference of the ACM Special Interest Group on Data Communication on the applications, technologies, architectures, and protocols for computer communication (SIGCOMM). 359–376. [42]Ji Lin, Jiaming Tang, Haotian Tang, Shang Yang, Wei-Ming Chen, Wei-Chen Wang, Guangxuan Xiao, Xingyu Dang, Chuang Gan, and Song Han. 2024. Awq: Activation-aware weight quantization for on-device llm compression and accel- eration. Proceedings of Machine Learning and Systems (MLsys) 6 (2024), 87–100. [43]Cihang Liu, Lan Zhang, Zongqian Liu, Kebin Liu, Xiangyang Li, and Yunhao Liu. 2016. Lasagna: Towards deep hierarchical understanding and searching over mobile sensing data. In Annual International Conference on Mobile Computing and Networking (MobiCom). 334–347. [44]Haibo Liu, Chen Gong, Zhenzhe Zheng, Shengzhong Liu, and Fan Wu. 2025. Enabling real-time inference in online continual learning via device-cloud collab- oration. In Proceedings of the ACM on Web Conference 2025 (W). 2043–2052. [45]Sicong Liu, Yingyan Lin, Zimu Zhou, Kaiming Nan, Hui Liu, and Junzhao Du. 2018. On-demand deep model compression for mobile devices: A usage-driven model selection framework. In International Conference on Mobile Systems, Applications, and Services (MobiSys). 389–400. [46]Sangkug Lym, Esha Choukse, Siavash Zangeneh, Wei Wen, Sujay Sanghavi, and Mattan Erez. 2019. Prunetrain: fast neural network training by dynamic sparse model reconfiguration. In International Conference for High Performance Computing, Networking, Storage and Analysis (SC). 1–13. [47]Xiaolong Ma, Fu-Ming Guo, Wei Niu, Xue Lin, Jian Tang, Kaisheng Ma, Bin Ren, and Yanzhi Wang. 2020. Pconv: The missing but desirable sparsity in dnn weight pruning for real-time execution on mobile devices. In Annual AAAI Conference on Artificial Intelligence (AAAI), Vol. 34. 5117–5124. [48] Silvano Martello and Paolo Toth. 1987. Algorithms for knapsack problems. North- Holland Mathematics Studies 132 (1987), 213–257. [49]Itzik Mazeh and Erez Shmueli. 2020. A personal data store approach for recom- mender systems: enhancing privacy without sacrificing accuracy. Expert Systems with Applications 139 (2020), 112858. [50]Abbas Mehrabi, Matti Siekkinen, and Antti Ylä-Jäski. 2018. Edge computing assisted adaptive mobile video streaming. IEEE Transactions on Mobile Computing (TMC) 18, 4 (2018), 787–800. SenSys ’26, May 11–14, 2026, Saint Malo, FranceGong et al. [51]Wei Niu, Jiexiong Guan, Yanzhi Wang, Gagan Agrawal, and Bin Ren. 2021. Dnnfu- sion: accelerating deep neural networks execution with advanced operator fusion. In International Conference on Programming Language Design and Implementation (PDLI). 883–898. [52]Wei Niu, Xiaolong Ma, Sheng Lin, Shihao Wang, Xuehai Qian, Xue Lin, Yanzhi Wang, and Bin Ren. 2020. Patdnn: Achieving real-time dnn execution on mobile devices with pattern-based weight pruning. In International Conference on Archi- tectural Support for Programming Languages and Operating Systems (ASPLOS). 907–922. [53]Nikola Obradovic, Aleksandar Kelec, and Igor Dujlovic. 2019. Performance analysis on Android SQLite database. In International Symposium INFOTEH- JAHORINA (INFOTEH). IEEE, 1–4. [54] Client AI Group of ByteDance. 2022. Introduction to ByteDance Pitaya. https: //xie.infoq.cn/article/d9a05a40ddbc1b01218f46a0a. [55]Gihwan Oh, Sangchul Kim, Sang-Won Lee, and Bongki Moon. 2015. SQLite optimization with phase change memory for mobile applications. Proceedings of the VLDB Endowment (VLDB) 8, 12 (2015), 1454–1465. [56] Felipe Pezoa, Juan L Reutter, Fernando Suarez, Martín Ugarte, and Domagoj Vrgoč. 2016. Foundations of JSON schema. In International Conference on World Wide Web (W). 263–273. [57]Ulrich Pferschy. 1999. Dynamic programming revisited: Improving knapsack algorithms. Computing 63, 4 (1999), 419–430. [58]J. Ross Quinlan. 1996. Learning decision tree classifiers. ACM Computing Surveys (CSUR) 28, 1 (1996), 71–72. [59]Hassan Ramchoun, Youssef Ghanou, Mohamed Ettaouil, and Mohammed Amine Janati Idrissi. 2016. Multilayer perceptron: Architecture optimization and training. (2016). [60] Harvey M Salkin and Cornelis A De Kluyver. 1975. The knapsack problem: a survey. Naval Research Logistics Quarterly 22, 1 (1975), 127–144. [61] Iqbal H Sarker, Mohammed Moshiul Hoque, Md Kafil Uddin, and Tawfeeq Al- sanoosy. 2021. Mobile data science and intelligent apps: concepts, AI-based modeling and research directions. Mobile Networks and Applications (2021), 285–303. [62]Shachar Schnapp and Sivan Sabato. 2021. Active feature selection for the mu- tual information criterion. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 9497–9504. [63] Leming Shen, Qiang Yang, Kaiyan Cui, Yuanqing Zheng, Xiao-Yong Wei, Jian- wei Liu, and Jinsong Han. 2024. Fedconv: A learning-on-model paradigm for heterogeneous federated clients. In International Conference on Mobile Systems, Applications, and Services (MobiSys). 398–411. [64] Alexander Shishkin, Anastasia Bezzubtseva, Alexey Drutsa, Ilia Shishkov, Eka- terina Gladkikh, Gleb Gusev, and Pavel Serdyukov. 2016. Efficient high-order interaction-aware feature selection based on conditional mutual information. Advances in neural information processing systems (NeurIPS) 29 (2016). [65]Xiaohu Tang, Yang Wang, Ting Cao, Li Lyna Zhang, Qi Chen, Deng Cai, Yunxin Liu, and Mao Yang. 2023. Lut-n: Empower efficient neural network inference with centroid learning and table lookup. In International Conference on Mobile Computing and Networking (MobiCom). 1–15. [66] Ikram Sumaiya Thaseen and Cherukuri Aswani Kumar. 2017. Intrusion detection model using fusion of chi-square feature selection and multi class SVM. Journal of King Saud University-Computer and Information Sciences 29, 4 (2017), 462–472. [67]Tuyen X Tran and Dario Pompili. 2018. Adaptive bitrate video caching and processing in mobile-edge computing networks. IEEE Transactions on Mobile Computing (TMC) 18, 9 (2018), 1965–1978. [68]Paul Voigt and Axel Von dem Bussche. 2017. The eu general data protection reg- ulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing 10, 3152676 (2017), 10–5555. [69]Manni Wang, Shaohua Ding, Ting Cao, Yunxin Liu, and Fengyuan Xu. 2021. Asymo: scalable and efficient deep-learning inference on asymmetric mobile cpus. In International Conference on Mobile Computing and Networking (MobiSys). 215–228. [70]Hao Wen, Yuanchun Li, Zunshuai Zhang, Shiqi Jiang, Xiaozhou Ye, Ye Ouyang, Yaqin Zhang, and Yunxin Liu. 2023. AdaptiveNet: Post-deployment Neural Archi- tecture Adaptation for Diverse Edge Environments. In International Conference on Mobile Computing and Networking (MobiCom). 28:1–28:17. [71] Wikipedia. 2025. TikTok. https://en.wikipedia.org/wiki/TikTok. [72] Wikipedia. 2025. Toutiao. https://en.wikipedia.org/wiki/Toutiao. [73]Sarah Wooders, Xiangxi Mo, Amit Narang, Kevin Lin, Ion Stoica, Joseph M Hellerstein, Natacha Crooks, and Joseph E Gonzalez. 2023. RALF: Accuracy- Aware Scheduling for Feature Store Maintenance. Proceedings of the VLDB Endowment (VLDB) 17, 3 (2023), 563–576. [74]Daliang Xu, Mengwei Xu, Qipeng Wang, Shangguang Wang, Yun Ma, Kang Huang, Gang Huang, Xin Jin, and Xuanzhe Liu. 2022. Mandheling: mixed- precision on-device DNN training with DSP offloading. In International Confer- ence on Mobile Computing and Networking (MobiCom). 214–227. [75]Mengwei Xu, Jiawei Liu, Yuanqiang Liu, Felix Xiaozhu Lin, Yunxin Liu, and Xuanzhe Liu. 2019. A First Look at Deep Learning Apps on Smartphones. In The ACM Web Conference (W). 2125–2136. [76]Hyunho Yeo, Youngmok Jung, Jaehong Kim, Jinwoo Shin, and Dongsu Han. 2018. Neural adaptive content-aware internet video delivery. In USENIX Symposium on Operating Systems Design and Implementation (OSDI). 645–661. [77] Mu Yuan, Lan Zhang, Fengxiang He, Xueting Tong, and Xiang-Yang Li. 2022. Infi: end-to-end learnable input filter for resource-efficient mobile-centric inference. In International Conference on Mobile Computing And Networking (MobiCom). 228–241. [78]Daochen Zha, Zaid Pervaiz Bhat, Kwei-Herng Lai, Fan Yang, Zhimeng Jiang, Shaochen Zhong, and Xia Hu. 2025. Data-centric artificial intelligence: A survey. ACM Computing Surveys (CSUR) 57, 5 (2025), 1–42. [79]Xiao Zhang, Changlin Mei, Degang Chen, Yanyan Yang, and Jinhai Li. 2019. Active incremental feature selection using a fuzzy-rough-set-based information entropy. IEEE Transactions on Fuzzy Systems 28, 5 (2019), 901–915. [80]Yongfeng Zhang, Xu Chen, Qingyao Ai, Liu Yang, and W Bruce Croft. 2018. Towards conversational search and recommendation: System ask, user respond. In ACM International Conference on Information and Knowledge Management (CIKM). 177–186. [81] Christos Ziakis, Maro Vlachopoulou, Theodosios Kyrkoudis, and Makrina Karagkiozidou. 2019. Important factors for improving Google search rank. Future Internet 11, 2 (2019), 32. Optimizing Feature Extraction for On-device Model Inference with User Behavior SequencesSenSys ’26, May 11–14, 2026, Saint Malo, France A Supplementary Dataset Analysis In this section, we include supplementary material to provide com- prehensive context for our industrial dataset, detailing the specific behavior traces of different video types across testing users. Diverse Video Behavior Traces. In the Evaluation Setup sec- tion, we detailed the overall behavior frequencies per 10-minute segment. Here, we break down those numbers by specific video type. In modern multimedia platforms like TikTok, behaviors such as watching normal short-form videos, live streams, shows, and creator homepage videos are all recorded as distinct behavior types. This distinction is necessary because these types involve heteroge- neous attribute set for description and target different audience. Figure 1 plot the frequencies of these distinct behaviors per 10- minute segment across our testing users. For the most popular short-form videos as shown in Figure 1(a), testing users consume an average of 4.02–6.15 short-form videos per 10 minutes at noon, 3.91–7.84 at evening, and 2.64–6.71 at night. Notably, the consump- tion rate is slightly lower at night, which may be explained by users dedicating more short, intense bursts of leisure time (like noon breaks) to this format. Shows and live streams are the next two most popular types and as shown in Figures 1(b) and 1(c), aver- age user consumptions per 10 minutes are 1.77-4.62 live stream and 4.04-5.15 shows at noon, 2.63-4.06 and 4.68-6.22 for evening as well as 1.50-4.37 and 2.72-7.05 for night. The lowest frequency behavior is viewing a creator’s homepage (Figure 1(d)), averaging only 0.52–2.40 views per 10 minutes across all three time periods. It is important to note that videos of different genres (e.g., comedy, drama) are typically recorded under the same behavior type. This is because they share the same underlying set of descriptive attributes, differing only in the values of those attributes. 1213 1819 21 2223 Time Segments across 3 Periods 0 5 10 15 20 Behavior Num. (per 10 min) P90 P80 P70 P60 P50 P30 Mean (a) Short-Form Video 1213 1819 21 2223 Time Segments across 3 Periods 0 4 8 12 16 Behavior Num. (per 10 min) P90 P80 P70 P60 P50 P30 Mean (b) Live Stream 1213 1819 21 2223 Time Segments across 3 Periods 0 6 12 18 24 Behavior Num. (per 10 min) P90 P80 P70 P60 P50 P30 Mean (c) Show 1213 1819 21 2223 Time Segments across 3 Periods 0 2 4 6 8 Behavior Num. (per 10 min) P90 P80 P70 P60 P50 P30 Mean (d) Homepage Figure 1: Behavior traces of testing users across three time periods (noon, evening and night). Each subfigure focuses on one video-related behavior and we plot the frequencies within each 10-minute segment of users with different activ- ity levels (P90, P80, P70, P60, P50, and P30 traces).