Skip to content

Latest commit

 

History

History
153 lines (135 loc) · 9.5 KB

File metadata and controls

153 lines (135 loc) · 9.5 KB

机器学习系统设计

ML design的核心,本质是训练一个model来实现某个任务,如prediction/ranking/classification

  • 建模design, 包括优化目标,feature,data,模型结构,评价标准等
  • 系统design, 偏重于在线serve大模型,包括feature store, ANN, ETL pipeline, MLOps等

例子

  • youtube recommendation/doordash search box/auto suggestion
  • design youtube violent content detection system
  • detecting unsafe content
  • design a monitoring system to realtime measure ML models, including features, score distribution, qps
  • abusive user detection

业务目标

  • improve engagement on a feed
  • improve customer churn
  • return items from search engine query

名词解释

  • 曝光(impression): 文档被用户看到
  • 点击率(click-through-rate,CTR): 文档d曝光的前提下,用户点击d的概率
  • 交互行为(engagement): 在点击的前提下, 文档的点赞、收藏、转发、关注作者、评论。电商的加购物车、下单、付款

面试过程

回答框架

  • 明确需求
    • 场景,功能,目标(engagement Or revenue),约束,如何转化为机器学习问题(如推荐转化为二分类模型和原因)
  • 数据
    • positive label and negative label
    • scale of the system, user和item有哪些数据和量级,一些可做特征的数据是否有log
    • 从2个方面identify data:training + label, testing + ground truth
    • label来源: 从交互中收集, 人工标注, 人工标注加无监督辅助, 增强数据
    • 数据探讨: bias, 非均衡, label质量
    • GDPR/privacy,数据脱敏,数据加密
    • train/test data和product上distribution不一样怎么办
    • data distribution随时间改变怎么办
  • 特征工程
    • user, item and cross, context
    • sparse and dense feature
    • 实际工作中,每个ML组都有自己不同的embedding set。互相使用别人的embedding set。怎么pre-train, fine-train, 怎么combine feature非常重要
    • feature的ABtest怎么做?不同traffic做
  • 模型
    • 模型选择,考虑系统方面的constraint。比如prediction latency, memory。怎么合理的牺牲模型的性能以换取constraint方面的benefit
    • 每个design的选择,像平时写design doc一样比较不同选项的优劣
    • 大多数场景,模型之外都需要额外的策略兜底
  • 评价
    • offline and online
    • 模型的评价,比如:点击,转化,是否有广告?考察的是GMV,还是转化订单?
  • 部署
    • server or device
    • all users or a part of users
    • statically, dynamically(server or device) or model streaming
  • serving
    • batch prediction or online prediction
  • monitoring
    • 监控latency,QPS,precision,recall等参数
  • maintain

过程

  • 一边白板画框图,一边告知面试官我要讲某几个部分
  • 整个过程,讲清楚主题之前,不要陷入任何一部分的细节挖掘。随着问题介绍,data和细节都会明确
  • 每个部分,尤其是自己熟悉的方面,要主动讲,因为每个部分都很重要
  • move之前最后确认:Is there anywhere that you feel I missed?

问答

  • how to scale

    • Scaling general SW system (distributed servers, load balancer, sharding, replication, caching, etc)
    • Train data / KB partitioning
    • Distributed ML
    • Data parallelism (for training)
    • Model parallelism (for training, inference)
    • Asynchronous SGD
    • Synchronous SGD
    • Distributed training
    • Data parallel DT, RPC based DT
    • Scaling data collection
    • machine translation for 1000 languages
      • NLLB
    • embedding-> Deep Hash Embedding
  • Monitoring, failure tolerance, updating (below)

  • Auto ML (soft: HP tuning, hard: arch search (NAS))

  • 线上线下不一致

  • 不同的数据用什么方式存储

  • data pipeline怎么设计

  • serving

    • Online A/B testing
      • Based on the online metrics we would select a significance level 𝛼 and power threshold 1 – 𝛽
      • Calculate the required sample size per variation
        • The required sample size depends on 𝛼, 𝛽, and the MDE Minimum Detectable Effect – the target relative minimum increase over the baseline that should be observed from a test
      • Randomly assign users into control and treatment groups (discuss with the interviewer whether we will split the candidates on the user level or the request level)
      • Measure and analyze results using the appropriate test. Also, we should ensure that the model does not have any biases.
    • If we are serving batch features they have to be handled offline and served at real time so we have to have daily/weekly jobs for generating this data.
    • If we are serving real time features then they need to be fetched/derived at request time and we need to be aware of scalability or latency issues (load balancing), we may need to create a feature store to lookup features at serve time and maybe some caching depending on the use case.
    • Where to run inference: if we run the model on the user’s phone/computer then it would use their memory/battery but latency would be quick, on the other hand, if we store the model on our own service we increase latency and privacy concerns but removes the burden of taking up memory and battery on the user’s device.
    • how often we would retrain the model. Some models need to be retrained every day, some every week and others monthly/yearly. Always discuss the pros and cons of the retraining regime you choose
  • deploy

    • model serving是典型的low latency high qps
    • 负载均衡和自动伸缩:
    • latency如何优化
    • 这么多server如何deploy,以及如何push新的model version,在更新的时候如何保证qps不degrade
  • Monitoring Performance

    • Latency (P99 latency every X minutes)
    • Biases and misuses of your model
    • Performance Drop
    • Data Drift
    • CPU load
    • Memory Usage

参考