今天15:00 | ICML-5 不见不散!

点击蓝字

21f2b6309eb382751ec9d02dca530ea7.png

关注我们

AI TIME欢迎每一位AI爱好者的加入!

fee0ff3bd8d972edf3a444c34553ec06.png

9月23日 15:00~20:30

AI TIME特别邀请了多位PhD,带来ICML-5!

f913cec4634295689af1355c20c88033.png

哔哩哔哩直播通道

扫码关注AITIME哔哩哔哩官方账号

观看直播

链接:https://live.bilibili.com/21813994

15:00-17:00

★ 嘉宾介绍 ★

343909842747ecf03d78e482bdcee64c.png

邓维建

澳洲国立大学二年级在读博士,在 Prof. Stephen Gould  和 Dr. Liang Zheng 指导下进行模型泛化能力研究。

个人主页https://weijiandeng.xyz。

报告题目:

4ff03caf6c4de9a533f0656e72772e38.gif

旋转预测能够告诉我们分类器

准确度的哪些信息?

内容简介:

Understanding classifier decision under novel environments is central to the community, and a common practice is evaluating it on labeled test sets. However, in real-world testing, image annotations are difficult and expensive to obtain, especially when the test environment is changing. A natural question then arises: given a trained classifier, can we evaluate its accuracy on varying unlabeled test sets? In this work, we train semantic classification and rotation prediction in a multi-task way. On a series of datasets, we report an interesting finding, i.e., the semantic classification accuracy exhibits a strong linear relationship with the accuracy of the rotation prediction task (Pearson’s Correlation r > 0.88). This finding allows us to utilize linear regression to estimate classifier performance from the accuracy of rotation prediction which can be obtained on the test set through the freely generated rotation labels.

b7c2db1a1b2f436d4b114d07252e6571.png

刘旭彤

香港中文大学在读博士生,师从John C.S. Lui, 研究兴趣包括强化学习、在线学习理论,特别是组合多臂老虎机(combinatorial multi-armed badnit)的算法设计、理论分析和在社交网络中的应用。

个人主页:https://xutongliu.me/.

报告题目:

9bba3b574a126bf989b124e7bd3db80f.gif

利用随机游走算法的多层

网络探索:从离线优化到在线学习

内容简介:

Multi-layered network exploration (MuLaNE) problem is an important problem abstracted from many applications. In MuLaNE, there are multiple network layers where each node has an importance weight and each layer is explored by a random walk. The MuLaNE task is to allocate total random walk budget B into each network layer so that the total weights of the unique nodes visited by random walks are maximized. We systematically study this problem from offline optimization to online learning. For the offline optimization setting where the network structure and node weights are known, we provide greedy based constant-ratio approximation algorithms for overlapping networks, and greedy or dynamic-programming based optimal solutions for non-overlapping networks. For the online learning setting, neither the network structure nor the node weights are known initially. We adapt the combinatorial multi-armed bandit framework and design algorithms to learn random walk related parameters and node weights while optimizing the budget allocation in multiple rounds, and prove that they achieve logarithmic regret bounds. Finally, we conduct experiments on a real-world social network dataset to validate our theoretical results.

7fce2084488e9f8fd329d6b4111c4b8d.png

滕佳烨

清华大学交叉信息研究院二年级博士生,导师为袁洋助理教授。主要研究方向是神经网络泛化理论,以及统计理论与机器学习的结合,包括因果推断、共形预测等。

报告题目:

950ff93d726c41f7bb45db88f8f61b8b.gif

T-SCI: 基于Cox-MLP模型的

二阶段共形预测

内容简介:

It is challenging to deal with censored data, where we only have access to the incomplete information of survival time instead of its exact value.

Fortunately, under linear predictor assumption, people can obtain guaranteed coverage for the confidence band of survival time using methods like Cox Regression. However, when relaxing the linear assumption with neural networks (e.g., Cox-MLP), we lose the guaranteed coverage. To recover the guaranteed coverage without linear assumption, we propose two algorithms based on conformal inference under strong ignorability assumption. In the first algorithm WCCI, we revisit weighted conformal inference and introduce a new non-conformity score based on partial likelihood. We then propose a two-stage algorithm T-SCI, where we run WCCI in the first stage and apply quantile conformal inference to calibrate the results in the second stage. Theoretical analysis shows that T-SCI returns guaranteed coverage under milder assumptions than WCCI. We conduct extensive experiments on synthetic data and real data using different methods, which validate our analysis.

bb9960fb39d76e27ab4b6cdf1a55dddb.png

谢肖飞

目前是新加坡南洋理工大学的校长博士后,曾毕业于天津大学。主要研究方向包括软件工程、安全以及人工智能领域,并在相关领域的顶级期刊与会议上发表论文多篇,包括 ICSE,ISSTA,FSE,ASE, TSE,TOSEM,IJCAI,ECCV,ICCV,ICLR, ICML,NeurIPS,CCS,AAAI,TIFS,TDSC等。其中那个两次获得 ACM SIGSOFT Distinguished Paper Awards (FSE'16, ASE'19).

报告题目:

ea7ff013b8d7cbd44d8671592b13d751.gif

基于模型分析的循环神经网络自动修复

内容简介:

Deep neural networks are vulnerable to adversarial attacks. Due to their black-box nature, it is rather challenging to interpret and properly repair these incorrect behaviors. This paper focuses on interpreting and repairing the incorrect behaviors of Recurrent Neural Networks (RNNs). We propose a lightweight model-based approach (RNNRepair) to help understand and repair incorrect behaviors of an RNN. Specifically, we build an influence model to characterize the stateful and statistical behaviors of an RNN over all the training data and to perform the influence analysis for the errors. Compared with the existing techniques on influence function, our method can efficiently estimate the influence of existing or newly added training samples for a given prediction at both sample level and segmentation level. Our empirical evaluation shows that the proposed influence model is able to extract accurate and understandable features. Based on the influence model, our proposed technique could effectively infer the influential instances from not only an entire testing sequence but also a segment within that sequence. Moreover, with the sample-level and segment-level influence relations, RNNRepair could further remediate two types of incorrect predictions at the sample level and segment level.

19:30-20:30

b1cebcc97db0c5c0a4f4556cfdc38ab5.png

李晓宇

波士顿大学系统工程博士四年级学生,导师是Dr. Francesco Orabona。研究方向主要包括随机优化,机器学习理论。

报告题目:

c87d7fceb196cd9285470cf9d3206d23.gif

随机梯度下降与指数步长及余弦步长

内容简介:

Stochastic Gradient Descent (SGD) is a popular tool in training large-scale machine learning models. Its performance, however, is highly variable, depending crucially on the choice of the step sizes. Accordingly, a variety of strategies for tuning the step sizes have been proposed, ranging from coordinate-wise approaches (a.k.a. “adaptive” step sizes) to sophisticated heuristics to change the step size in each iteration. In this paper, we study two step size schedules whose power has been repeatedly confirmed in practice: the exponential and the cosine step sizes. For the first time, we provide theoretical support for them proving convergence rates for smooth non-convex functions, with and without the Polyak-Łojasiewicz (PL) condition. Moreover, we show the surprising property that these two strategies are adaptive to the noise level in the stochastic gradients of PL functions. That is, contrary to polynomial step sizes, they achieve almost optimal performance without needing to know the noise level nor tuning their hyperparameters based on it. Finally, we conduct a fair and comprehensive empirical evaluation of real-world datasets with deep learning architectures. Results show that, even if only requiring at most two hyperparameters to tune, these two strategies best or match the performance of various finely-tuned state-of-the-art strategies.

c2d37a01ccbf4ea3a251b299805aa7aa.png

郭文博

宾州州立大学博士,研究方向主要是机器学习和安全, 并在相关领域的顶级期刊与会议上发表论文多篇,包括 ICML,NeurIPS,CCS,USENIX Security, NDSS等。获得 ACM CCS Outstanding Paper Awards, 2018, IBM PhD. Fellowship Award, 2020, 以及Facebook/Baidu Ph.D. Fellowship finalists, 2020.

报告题目:

187c8fd85ad4fc79740f13aefba06544.gif

对抗深度强化学习策略的训练和理论保障

内容简介:

In a two-player deep reinforcement learning task, recent work shows an attacker could learn an adversarial policy that triggers a target agent to perform poorly and even react in an undesired way. However, its efficacy heavily relies upon the zero-sum assumption made in the two-player game. In this work, we propose a new adversarial learning algorithm. It addresses the problem by resetting the optimization goal in the learning process and designing a new surrogate optimization function. Our experiments show that our method significantly improves adversarial agents’ exploitability compared with the state-of-art attack. Besides, we also discover that our method could augment an agent with the ability to abuse the target game’s unfairness. Finally, we show that agents adversarially retrained against our adversarial agents could obtain stronger adversary-resistance.

# 今日视推荐 #

直播结束后我们会邀请讲者在微信群中与大家答疑交流,请添加“AI TIME小助手(微信号:AITIME_HY)”,回复“icml”,将拉您进“AI TIME ICML 会议交流群”!

ef4d1630e0e808b80cda2db6ad834709.gif

AI TIME微信小助手

b741f4feb399cc21d7fe4d6f7440cbcc.png

主       办:AI TIME 

合作媒体:学术头条、AI 数据派

合作伙伴:智谱·AI、中国工程院知领直播、学堂在线、学术头条、biendata、 Ever链动

AI TIME欢迎AI领域学者投稿,期待大家剖析学科历史发展和前沿技术。针对热门话题,我们将邀请专家一起论道。同时,我们也长期招募优质的撰稿人,顶级的平台需要顶级的你,

请将简历等信息发至yun.he@aminer.cn!

微信联系:AITIME_HY

AI TIME是清华大学计算机系一群关注人工智能发展,并有思想情怀的青年学者们创办的圈子,旨在发扬科学思辨精神,邀请各界人士对人工智能理论、算法、场景、应用的本质问题进行探索,加强思想碰撞,打造一个知识分享的聚集地。

6b7b916399be7022c6ee8de8f232940b.png

更多资讯请扫码关注

a521c07dc9cce9442cafaa3c3de0e090.png

我知道你“在看”哟~

764a6684ac66e1e175f696919f3135b6.gif

点击 阅读原文 了解更多精彩

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值