• ICML-20 待读的 Paper


    2020.9.12
    花了一上午的时间,过了一遍 ICML-2020 Accepted Paper List, 挑出了自己想读的 Paper。
    主要关注于自己的一些研究点。

    Noisy Labels

    • Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels
    • SIGUA: Forgetting May Make Learning with Noisy Labels More Robust
    • Error-Bounded Correction of Noisy Labels
    • Does label smoothing mitigate label noise?
    • Deep k-NN for Noisy Labels
    • Improving generalization by controlling label-noise information in neural network weights
    • Normalized Loss Functions for Deep Learning with Noisy Labels
    • Variational Label Enhancement
    • Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
    • Searching to Exploit Memorization Effect in Learning with Noisy Labels
    • Label-Noise Robust Domain Adaptation
    • Training Binary Neural Networks through Learning with Noisy Supervision
    • Learning with Bounded Instance- and Label-dependent Label Noise
    • Progressive Identification of True Labels for Partial-Label Learning
    • Learning with Multiple Complementary Labels
    • Unbiased Risk Estimators Can Mislead: A Case Study of Learning with Complementary Labels

    Semi-Supervised Learning

    • Semi-Supervised Learning with Normalizing Flows
    • Negative Sampling in Semi-Supervised learning
    • Poisson Learning: Graph Based Semi-Supervised Learning At Very Low Label Rates
    • Time-Consistent Self-Supervision for Semi-Supervised Learning
    • Safe Deep Semi-Supervised Learning for Unseen-Class Unlabeled Data
    • Deep Streaming Label Learning

    Domain Adaptation

    • Continuously Indexed Domain Adaptation
    • RIFLE: Backpropagation in Depth for Deep Transfer Learning through Re-Initializing the Fully-connected LayEr
    • Implicit Class-Conditioned Domain Alignment for Unsupervised Domain Adaptation
    • Understanding Self-Training for Gradual Domain Adaptation
    • Do We Really Need to Access the Source Data? Source Hypothesis Transfer for Unsupervised Domain Adaptation
    • LTF: A Label Transformation Framework for Correcting Label Shift
    • Label-Noise Robust Domain Adaptation

    Data Bias, Weighting

    • Adversarial Filters of Dataset Biases
    • Optimizing Data Usage via Differentiable Rewards
    • Data preprocessing to mitigate bias: A maximum entropy based approach
    • DeBayes: a Bayesian Method for Debiasing Network Embeddings
    • A Distributional Framework For Data Valuation

    Class-Imbalance

    • Class-Weighted Classification: Trade-offs and Robust Approaches
    • Online Continual Learning from Imbalanced Data
    • Logistic Regression for Massive Data with Rare Events

    MixUp, Interpolation, Extrapolation, etc.

    • Puzzle Mix: Exploiting Saliency and Local Statistics for Optimal Mixup
    • Learning Representations that Support Extrapolation
    • Extrapolation for Large-batch Training in Deep Learning
    • Training Neural Networks for and by Interpolation
    • Sequence Generation with Mixed Representations

    PU Learning

    • Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training

    Active Learning

    • Adaptive Region-Based Active Learning

    GCN or Recommendation System

    • Continuous Graph Neural Networks
    • Simple and Deep Graph Convolutional Networks
    • Graph Convolutional Network for Recommendation with Low-pass Collaborative Filters
    • Generalization and Representational Limits of Graph Neural Networks
    • Graph-based Nearest Neighbor Search: From Practice to Theory
    • Ordinal Non-negative Matrix Factorization for Recommendation
    • Improved Communication Cost in Distributed PageRank Computation – A Theoretical Study
    • Optimization and Analysis of the pAp@k Metric for Recommender Systems
    • Scalable and Efficient Comparison-based Search without Features
    • Learning to Rank Learning Curves
    • When Does Self-Supervision Help Graph Convolutional Networks?

    Neural ODE

    • Towards Adaptive Residual Network Training: A Neural-ODE Perspective
    • Combining Differentiable PDE Solvers and Graph Neural Networks for Fluid Flow Prediction
    • Approximation Capabilities of Neural ODEs and Invertible Residual Networks

    Calibration, Confidence, Out-of-distribution

    • Confidence-Aware Learning for Deep Neural Networks
    • Being Bayesian, Even Just a Bit, Fixes Overconfidence in ReLU Networks
    • SDE-Net: Equipping Deep Neural Networks with Uncertainty Estimates
    • Detecting Out-of-Distribution Examples with Gram Matrices
    • Interpretable, Multidimensional, Multimodal Anomaly Detection with Negative Sampling for Detection of Device Failure
    • Uncertainty Estimation Using a Single Deep Deterministic Neural Network
    • How Good is the Bayes Posterior in Deep Neural Networks Really?

    Federated Learning, Fairness

    • Fair k-Centers via Maximum Matching
    • Federated Learning with Only Positive Labels

    Interesting Problems, Settings

    • Why Are Learned Indexes So Effective?
    • Learning with Feature and Distribution Evolvable Streams
    • Cost-effectively Identifying Causal Effects When Only Response Variable is Observable
    • Rigging the Lottery: Making All Tickets Win
    • Do We Need Zero Training Loss After Achieving Zero Training Error?
    • Small Data, Big Decisions: Model Selection in the Small-Data Regime
    • Why bigger is not always better: on finite and infinite neural networks
    • On Learning Sets of Symmetric Elements
    • Collaborative Machine Learning with Incentive-Aware Model Rewards
    • Generalisation error in learning with random features and the hidden manifold model
    • Sample Amplification: Increasing Dataset Size even when Learning is Impossible
    • When are Non-Parametric Methods Robust?
    • Performative Prediction
    • Supervised learning: no loss no cry
    • Teaching with Limited Information on the Learner's Behaviour
    • Learning De-biased Representations with Biased Representations
    • Do RNN and LSTM have Long Memory?
    • It's Not What Machines Can Learn, It's What We Cannot Teach
    • Enhancing Simple Models by Exploiting What They Already Know

    Interesting Theory

    • On the Generalization Benefit of Noise in Stochastic Gradient Descent
    • Let's Agree to Agree: Neural Networks Share Classification Order on Real Datasets
    • Optimizer Benchmarking Needs to Account for Hyperparameter Tuning
    • On the Noisy Gradient Descent that Generalizes as SGD
    • Rethinking Bias-Variance Trade-off for Generalization of Neural Networks
    • Understanding and Mitigating the Tradeoff between Robustness and Accuracy
    • The Implicit and Explicit Regularization Effects of Dropout
    • Optimal Continual Learning has Perfect Memory and is NP-hard
    • Curvature-corrected learning dynamics in deep neural networks
    • Explainable k-Means and k-Medians Clustering
    • Near-Tight Margin-Based Generalization Bounds for Support Vector Machines
    • Uniform Convergence of Rank-weighted Learning
    • Decision Trees for Decision-Making under the Predict-then-Optimize Framework

    Interesting Algorithm

    • SoftSort: A Continuous Relaxation for the argsort Operator
    • Boosting Deep Neural Network Efficiency with Dual-Module Inference
    • Circuit-Based Intrinsic Methods to Detect Overfitting
    • Learning Similarity Metrics for Numerical Simulations
    • Deep Divergence Learning
    • Consistent Estimators for Learning to Defer to an Expert
    • Smaller, more accurate regression forests using tree alternating optimization
    • Learning To Stop While Learning To Predict
    • DROCC: Deep Robust One-Class Classification

    Point Cloud, 3 dimension

    • Hypernetwork approach to generating point clouds
  • 相关阅读:
    查看Linux系统版本信息
    ensemble github强大的下载安装功能--ensembl-git-tools
    Linux系统非root用户安装perl模块
    Linux下安装与使用本地的perl模块
    MATLAB 2014a (8.3) Compiler Runtime (MCR)
    GEO--工具 ScanGEO
    vcf2maf
    RNAseq 流程
    pathway一些网站
    abbitMQ整合Spring Booot【点对点模式】
  • 原文地址:https://www.cnblogs.com/Gelthin2017/p/13652981.html
Copyright © 2020-2023  润新知