楼主: wwqqer
20632 196

【阿尔法系列】美知名对冲基金介绍深度学习(Deep Learning)最新研究成果!   [推广有奖]

回帖奖励 29 个论坛币 回复本帖可获得 1 个论坛币奖励! 每人限 1 次(中奖概率 30%)

版主

泰斗

65%

还不是VIP/贵宾

-

TA的文库  其他...

Wiley文库

Springer文库

全球著名CRC出版社文库

威望
17
论坛币
109170 个
通用积分
102440.6899
学术水平
5957 点
热心指数
6460 点
信用等级
5272 点
经验
3927 点
帖子
7502
精华
93
在线时间
9406 小时
注册时间
2007-12-10
最后登录
2024-4-18

二级伯乐勋章 一级伯乐勋章 初级学术勋章 中级学术勋章 初级热心勋章 中级热心勋章 初级信用勋章 中级信用勋章 高级学术勋章 高级热心勋章 特级学术勋章 高级信用勋章 特级信用勋章 特级热心勋章

楼主
wwqqer 在职认证  发表于 2016-12-25 08:51:51 |只看作者 |坛友微信交流群|倒序 |AI写论文
相似文件 换一批

+2 论坛币
k人 参与回答

经管之家送您一份

应届毕业生专属福利!

求职就业群
赵安豆老师微信:zhaoandou666

经管之家联合CDA

送您一个全额奖学金名额~ !

感谢您参与论坛问题回答

经管之家送您两个论坛币!

+2 论坛币


【阿尔法系列】将定期介绍业界最新成果及各种阿尔法策略!想要随时跟踪【阿尔法系列】,请点击头像下方“加关注”。
关注成功后,查看这里即可三步走把千本好书“一网打尽”!

[相关阅读]
【阿尔法系列】(资料汇总帖,附链接,持续添加中)
【机器学习系列】(资料汇总帖,附链接,持续添加中)

【经典教材系列】深度学习 Deep Learning (2016,Goofllow)
【经典教材系列】加强学习入门 Reinforcement Learning: An Introduction 第二版

人工智能入门: 灵魂机器的时代:当计算机超过人类智能 The Age of Spiritual Machines
人工智能入门: 奇点临近 The Singularity Is Near: When Humans Transcend Biology
人工智能入门: 机器学习 Machine Learning (Tom Michell)

人工智能入门: 人工智能的未来 On Intelligence
人工智能入门: 人工智能:智能系统指南 (第三版)Artificial Intelligence: A Guide

人工智能入门: Artificial Intelligence: A Modern Approach (第3版,高清)
人工智能入门: Artificial Intelligence: Structures and Strategies for Complex Problem (第6版)

【畅销书系列】Rise of the Robots: Technology and the Threat of a Jobless Future
【畅销书系列】Humans Need Not Apply: A Guide to Wealth and Work in the Age of
【畅销书系列】The Second Machine Age: Work, Progress, and Prosperity in a


今年(2016年6月)的国际机器学习大会(the International Conference on Machine Learning,ICML)的主题是深度学习(Deep Learning)。本次大会共分四个主题:Recurrent Neural Networks, Unsupervised Learning,Supervised Training Methods,Deep Reinforcement Learning。附件中的论文分别代表着这几研究方向的最新前沿。

深度学习(Deep Learning)是机器学习的一个分支,至今已有数种深度学习框架,如深度神经网络、卷积神经网络,深度信念网络和递归神经网络等。有些已被应用于计算机视觉、语音识别、自然语言处理、音频识别与生物信息学等领域并取得了极好的效果。

值得一提的是,2016年3月谷歌子公司DeepMind开发的围棋人工智能程序“AlphaGo
正是运用了“Deep Reinforcement Learning技术 (https://deepmind.com/blog/deep-reinforcement-learning/),在与韩国李世石九段围棋五番棋大战中,历史上首次击败人类围棋顶尖高手。由于围棋的复杂程度远远超过其他任何游戏,人们之前一直认为机器想要战胜人类还遥遥无期。在人机五番棋大战中,人们的态度从起先轻视,怀疑,再到惊讶,迷茫,最后绝望,醒悟,不经意间经历了一次心灵上极其震撼的洗礼。AlphaGo的胜利意味着人工智能的无限可能,这是一件可以写入人类历史的里程碑事件!

与此同时在量化投资领域,人们也在探索运用机器学习技术的可能性。许多对冲基金(例如,Man Group, Two Sigma, DE Shaw,等等)都投入了大量的人力和财力以求占得先机。本文是美国知名对冲基金Two Sigma的专家Vinod Valsalam对本次国际机器学习大会的总结(论文在附件中)。可以想象在不远的将来,机器学习将成为量化投资中的制胜法宝!

地址回复可见,17篇最新论文,降价出售3

本帖隐藏的内容

paper.zip (17.91 MB, 需要: 23 个论坛币)




photo.php.jpg



Machine learning offers powerful techniques to find patterns in data for solving challenging predictive problems. The dominant track at the International Conference onMachine Learning (ICML) in New York this year was deep learning, which uses artificial neural networks to solve problems by learning feature representations from large amounts of data.

Significant recent successes in applications such as image and speech recognition, and natural language processing, have helped fuel an explosion of interest in deep learning. And new research in the field is continuing to push the boundaries of applications, techniques, and theory. Below, Two Sigma research scientist Vinod Valsalam provides an overview of some of the most interesting research presented at ICML 2016, covering recurrent neural networks, unsupervised learning, supervised training methods, and deep reinforcement methods.

1. Recurrent Neural Networks

Unlike feed-forward networks, the outputs of recurrent neural networks (RNNs) can depend on past inputs, providing a natural framework for learning from time series and sequential data. But training them for tasks that require long-term memory is especially difficult due to the vanishing and exploding gradients problem, i.e., the error signals for adapting network weights become increasingly difficult to propagate through the network. Specialized network architectures such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) mitigate this problem by utilizing gating units, a technique that has been very successful in tasks such as speech recognition and language modeling. An alternative approach that is now gaining more focus is to constrain the weight matrices in a way that is more conducive to gradient propagation, as explored in the following papers.

Unitary Evolution Recurrent Neural Networks
Arjovsky, M., Shah, A., & Bengio, Y. (2016)
The problem of vanishing and exploding gradients occurs when the magnitude of the eigenvalues of weight matrices deviate from 1. Therefore, the authors use weight matrices that are unitary to guarantee that the eigenvalues have magnitude 1. The challenge with this constraint is to ensure that the matrices remain unitary when updating them during training without performing excessive computations. Their strategy is to decompose each unitary weight matrix into the product of several simple unitary matrices. The resulting parameterization makes it possible to learn the weights efficiently while providing sufficient expressiveness. They demonstrate state of the art performance on standard benchmark problems such as the copy and addition tasks. An additional benefit of their approach is that it is relatively insensitive to parameter initialization, since unitary matrices preserve norms.

Recurrent Orthogonal Networks and Long-Memory Tasks
Henaff, M., Szlam, A., & LeCun, Y. (2016)
In this paper, the authors construct explicit solutions based on orthogonal weight matrices for the copy and addition benchmark tasks. Orthogonal matrices avoid the vanishing and exploding gradients problem in the same way as unitary matrices, but they have real-valued entries instead of complex-valued entries. The authors show that their hand-designed networks work well when applied to the task for which they are designed, but produce poor results when applied to other tasks. These experiments illustrate the difficulty of designing general networks that perform well on a range of tasks.

Strongly-Typed Recurrent Neural Networks
Balduzzi, D., & Ghifary, M. (2016)
Physics has the notion of dimensional homogeneity, i.e. it is only meaningful to add quantities of the same physical units. Types in programming languages express a similar idea. The authors extend these ideas to constrain RNN design. They define a type as an inner product space with an orthonormal basis. The operations and transformations that a neural network performs can then be expressed in terms of types. For example, applying an activation function to a vector preserves its type. In contrast, applying an orthogonal weight matrix to a vector transforms its type. The authors argue that the feedback loop of RNNs produces vectors that are type-inconsistent with the feed-forward vectors for addition. While symmetric weight matrices are one way to preserve types in feedback loops, the authors tweak the LSTM and GRU networks to produce variants that have strong types. Experiments were inconclusive in showing better generalization of typed networks, but they are an interesting avenue for further research.

2. Unsupervised Learning

The resurgence of deep learning in the mid-2000s was made possible to a large extent by using unsupervised learning to pre-train deep neural networks to establish good initial weights for later supervised training. Later, using large labeled data sets for supervised training was found to obviate the need for unsupervised pre-training. But more recently, there has been renewed interest in utilizing unsupervised learning to improve the performance of supervised training, particularly by combining both into the same training phase.

Augmenting Supervised Neural Networks with Unsupervised Objectives for Large-scale Image Classification
Zhang, Y., Lee, K., & Lee, H. (2016)
This paper starts out with a brief history of using unsupervised and semi-supervised methods in deep learning. The authors showed how such methods can be scaled to solve large-scale problems. Using their approach, existing neural network architectures for image classification can be augmented with unsupervised decoding pathways for image reconstruction. The decoding pathways consist of a deconvolutional network that mirrors the original network using autoencoders. They initialized the weights for the encoding pathway with the original network and for the decoding pathway with random values. Initially, they trained only the decoding pathway while keeping the encoding pathway fixed. Then they fine-tuned the full network with a reduced learning rate. Applying this method to a state-of-the-art image classification network boosted its
performance significantly.

Deconstructing the Ladder Network Architecture
Pezeshki, M., Fan, L., Brakel, P., Courville, A., & Bengio, Y. (2016)
A different approach for combining supervised and unsupervised training of deep neural networks is the Ladder Network architecture. It also improves the performance of an existing classifier network by augmenting it with an auxiliary decoder network, but it has additional lateral connections between the original and decoder networks. The resultant network forms a deep stack of denoising autoencoders that is trained to reconstruct each layer from a noisy version. In this paper, the authors studied the ladder architecture systematically by removing its components one at a time to see how much each component contributed to performance. They found that the lateral connections are the most important, followed by the injection of noise, and finally by the choice of the combinator function that combines the vertical and lateral connections. They also introduced a new combinator function that improved the already impressive performance of the ladder network on the Permutation-Invariant MNIST handwritten digit recognition task, both for the supervised and semi-supervised settings.



【阿尔法系列】将定期介绍业界最新成果及各种阿尔法策略!想要随时跟踪【阿尔法系列】,请点击头像下方“加关注”。关注成功后,查看这里即可三步走把千本好书“一网打尽”!


请继续往下看。。。。


二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

关键词:Learning earning Learn 深度学习 研究成果 人工智能 阿尔法 对冲基金 计算机 经典

已有 4 人评分经验 论坛币 学术水平 热心指数 信用等级 收起 理由
oliyiyi + 100 精彩帖子
h2h2 + 5 + 5 + 5 精彩帖子
accumulation + 100 + 100 + 1 + 1 + 1 精彩帖子
Nicolle + 100 + 1 精彩帖子

总评分: 经验 + 300  论坛币 + 100  学术水平 + 7  热心指数 + 6  信用等级 + 6   查看全部评分

本帖被以下文库推荐

沙发
wwqqer 在职认证  发表于 2016-12-25 09:00:35 |只看作者 |坛友微信交流群

3. Supervised Training Methods

Historically, deep neural networks were known to be difficult to train using standard random initialization and gradient decent. However, new algorithms for initializing and training deep neural networks proposed in the last decade have produced remarkable successes. Research continues in this area to better understand existing training methods and to improve them.

Dropout distillation
Bulò, Samuel Rota, Porzi, L., & Kontschieder, P. (2016)
Dropout is a regularization technique that was proposed to prevent neural networks from overfitting. It drops units from the network randomly during training by setting their outputs to zero, thus reducing co-adaptation of the units. This procedure implicitly trains an ensemble of exponentially many smaller networks sharing the same parametrization. The predictions of these networks must then be averaged at test time, which is unfortunately intractable to compute precisely. But the averaging can be approximated by scaling the weights of a single network.

However, this approximation may not produce sufficient accuracy in all cases. The authors introduce a better approximation method called dropout distillation that finds a predictor with minimal divergence from the ideal predictor by applying stochastic gradient descent. The distillation procedure can even be applied to networks already trained using dropout by utilizing unlabeled data. Their results on benchmark problems show consistent improvements over standard dropout.

Normalization Propagation: A Parametric Technique for Removing Internal Covariate Shift in Deep Networks
Arpit, D., Zhou, Y., Kota, B., & Govindaraju, V. (2016)
One of the difficulties of training deep neural networks is that the distribution of input activations to each hidden layer may shift during training. One way to address this problem, known as internal covariate shift, is to normalize the input activations to each hidden layer using the Batch Normalization (BN) technique. However, BN has a couple of drawbacks: (1) its estimates of mean and standard deviation of input activations are inaccurate, especially during initial iterations, because they are based on mini-batches of training data and (2) it cannot be used with batch-size of one. To address these drawbacks, the authors introduce normalization propagation, which is based on a data-independent closed-form estimate of mean and standard deviation for every layer. It is based on the observation that the pre-activation values of ReLUs in deep networks follow a Gaussian distribution. The normalization property can then be forward-propagated to all hidden layers during training. The authors show that their method achieves better convergence stability than BN during training. It is also faster because it doesn't have to compute a running estimate of the mean and standard deviation of
the hidden layer activations.

Scalable Gradient-Based Tuning of Continuous Regularization Hyperparameters
Luketina, J., Raiko, T., Berglund, M., & Greff, K. (2016)
Tuning hyperparameters is often necessary to get good results with deep neural networks. Typically, the turning is performed either by manual trial-and-error, by using search, or by evaluating validation set performance. The authors propose a gradient based method that is less tedious and less computationally expensive to find good regularization hyperparameters.

Unlike previous methods, their method is simpler and computationally lightweight, and it updates both hyperparameters and regular parameters using stochastic gradient descent in the same training run. The gradient of the hyperparameters is obtained from the cost of the unregularized model on the validation set. Although the authors show that their method is effective in finding good regularization hyperparameters, they haven't extended it to common training techniques such as dropout regularization and learning rate adaptation.

4. Deep Reinforcement Learning

The researchers at DeepMind extended the breakthrough successes of deep learning in supervised tasks to the challenging reinforcement learning domain of playing Atari 2600 games. Their basic idea was to leverage the demonstrated ability of deep learning to extract high-level features from raw high-dimensional data by training a deep convolutional network. However, reinforcement learning tasks such as playing games do not come with training data that are labeled with the correct move for each turn.

Instead, they are characterized by sparse, noisy, and delayed reward signals. Furthermore, training data are typically correlated and non-stationary. They overcame these challenges using stochastic gradient descent and experience replay to stabilize learning, essentially jump-starting the field of deep reinforcement learning.

Asynchronous Methods for Deep Reinforcement Learning
Mnih, V., Badia, A. P., Mirza, M., Graves, A., Lillicrap, T., Harley, T., Silver, D., Kavukcuoglu (2016)
The experience replay technique stabilizes learning by making it possible to batch or sample the training data randomly. However, it requires more memory and computation and applies only to off-policy learning algorithms such as Q-learning. In this paper, the authors introduce a new method based on asynchronously executing multiple agents on different instances of the environment. The resulting parallel algorithm effectively de-correlates the training data and makes it more stationary. Moreover, it makes it possible to extend deep learning to off-policy reinforcement learning algorithms such as SARSA and actor-critic methods. Their method combined with the actor-critic algorithm improved upon previous results on the Atari domain using
much less computation resources.

Dueling Network Architectures for Deep Reinforcement Learning
Wang, Z., Schaul, T., Hessel, M., van Hasselt, Hado, Lanctot, M., & de Freitas, Nando (2016)
This work, which won the Best Paper  award, introduces a new neural network architecture that complements the algorithmic advances in deep Q-learning networks (DQN) and experience replay. The authors point out that the value of an action choice from a given state need to be estimated only if that action has a consequence on what happens. The dueling network architecture leverages this observation by inserting two parallel streams of fully connected layers after the final convolutional layer of a regular DQN. One of the two streams estimates the state-value function while the other stream estimates the state-dependent advantage of taking an action. The output module of the network combines the activations of these two streams to produce the Q-values for each action. This architecture learns state-value functions more efficiently and produces better policy evaluations when actions have similar values or the number of actions is large.

Opponent Modeling in Deep Reinforcement Learning
He, H., Boyd-Graber, J., Kwok, K., & III, Hal Daumé (2016)
The authors introduce an extension of the deep Q-network (DQN) called Deep Reinforcement Opponent Network (DRON) for multi-agent settings, where the action outcome of the agent being controlled depends on the actions of the other agents (opponents). If the opponents use fixed policies, then standard Q-learning is sufficient.

However, opponents with non-stationary policies occur when they learn and adapt their strategies over time. In this scenario, treating the opponents as part of the world in a standard Q-learning setup masks changes in opponent behavior. Therefore, the joint policy of opponents must be taken into consideration when defining the Q-function. The DRON architecture implements this idea by employing an opponent network to learn opponent policies and a Q-network to evaluate actions for a state. The outputs of the two networks are combined using a Mixture-of-Experts network [13] to obtain the expected Q-value. DRON out-performed DQN in simulated soccer and a trivia game by discovering different strategy patterns of opponents.

Conclusions

Deep learning is experiencing a phase of rapid growth due to its strong performance in a number of domains, producing state of the art results and winning machine learning competitions. However, these successes have also contributed to a fair amount of hype. The papers presented at ICML 2016 provided an unvarnished view of a vibrant field in which researchers are working actively to overcome challenges in making deep learning techniques more powerful, and in extending their successes to other domains and larger problems.



使用道具

藤椅
wwqqer 在职认证  发表于 2016-12-25 09:01:22 |只看作者 |坛友微信交流群
想要随时跟踪最新好书,请点击头像下方“加关注”。关注成功后,查看这里即可:三步走把千本好书“一网打尽”!
欢迎订阅wwqqer文库!
[原创] 浅析动量因子(附带Matlab/SAS程序及经典文献85篇,免费
[原创] 如何复制对冲基金的成功?(hedge fund replication,附免费文献)
[原创] 对于目前流行的量化投资与smart beta策略的一些看法 (附免费文献10篇)
[原创] 庄子“逍遥”之我见

【经典教材系列】(资料汇总帖,附链接,持续添加中)
【金融教材系列】(资料汇总帖,附链接,持续添加中)
【统计教材系列】(资料汇总帖,附链接,持续添加中)
【大数据系列】(资料汇总帖,附链接,持续添加中)
【程序软件系列】(资料汇总帖,附链接,持续添加中)

【阿尔法系列】(资料汇总帖,附链接,持续添加中)
【大师系列】(资料汇总帖,持续添加中)
【华尔街系列】(资料汇总帖,附链接,持续添加中)
【Wiley应用量化金融系列】(资料汇总帖,附链接,持续添加中)
【Wiley-Kolb金融系列】(资料汇总帖,附链接,持续添加中)

【国际政经系列】(资料汇总帖,附链接,持续添加中)
【2008金融危机必读系列】(资料汇总帖,附链接,持续添加中)
【畅销书系列】(资料汇总帖,附链接,持续添加中)
【查理芒格系列】Charlie Munger 推荐的20本书!(附链接)
【西蒙系列】跨学科旅行家: 赫伯特 西蒙 (Herbert Simon)资料汇总帖
【弗格森系列】学术界里的明星与怪伽: 尼尔•弗格森(Niall Ferguson)著作汇总帖

美国《时代》杂志2015年度十大最佳非小说类作品(附链接)
《经济学 人》2015年度最佳书单(附链接)
《经济学 人》2014年度最佳书单(附链接)
亚马逊2015年度最佳商业投资类图书(附链接)
亚马逊2014年度最佳商业投资类图书(附链接)
2016年度英国《金融 时报》最佳商业图书书单(附链接)
2015年度英国《金融 时报》最佳商业图书书单(附链接)
2014年度英国《金融 时报》最佳商业图书书单(附链接)

2015年欧美政商学界精英的精彩阅读瞬间!
扎克伯格的读书年(A Year of Books)
比尔·盖茨2015年度推荐书单:关注事物的工作原理 (附链接)
比尔·盖茨2016年度推荐书单(附链接)
【独家发布】比尔·盖茨推荐的九本书----希望有人能将它们(感谢olderp的热心帮助)
比尔·盖茨最喜欢的商业书籍 (Bill Gates's Favorite Business Book)
【资源典藏】最值得收藏的创业书单:21本必读国外经典经管书籍都在这里了(感谢iRolly的热心帮助)

金融危机畅销书作家Peter Schiff系列
2015年光棍节推荐书单(附链接)
2015年,梁小民读了328本书,但只推荐这10本(感谢版主的热心帮助)
2015年最值得馆藏的20本商业图书(感谢chenyi112982的热心帮助)
美国知名财经作家Jason Zweig投资入门书推荐!(附链接)
史上最好的20条投资建议 The Best Investment Advice Of All Time (附链接)
资深业内人士推荐的10本交易书(附链接)Top Ten Trading Books I Have Read

[专题系列]
大牛Paul Krugman:日本,对不起!
[专题系列] Barra模型-RiskMetrics (RMA)-PMA资料(持续更新)
[专题系列] 主动投资与被动投资(active vs. passive),到底哪个更厉害?(免费!)
[专题系列] 行为经济学 From “Economic Man” to Behavioral Economics
[专题系列] 特朗普当选美国总统创造奇迹,我们见证历史!经典之作 Trump: The Art of the Deal

[专题系列] 斯坦福大学经济系是如何后来居上的?
[专题系列] ECB 终于把名义利率降为负值了!(附重要文献11篇,免费)
[专题系列] Frameworks for Central Banking in the Next Century(最新文献9篇,免费)
[专题系列] Energy Derivatives Pricing (能源衍生品定价介绍,27篇文献,全部免费)
[专题系列] 揭秘世界知名对冲基金AQR制胜交易策略!附带29篇文献

[专题系列] 有效市场假设(Efficient Market Hypothesis) :一场伟大的分歧!
[专题系列] 金融危机后,通胀目标(Inflation Targeting)是否仍然可行?
[专题系列] 非常规货币政策退出策略(Exit Strategy) 权威报告!
[专题系列] 回测过程中的过度拟合问题 (backtest overfitting,附最新文献2篇)
[专题系列] 做计量的朋友们,你们的标准误差(standard error)算对了吗?(附程序)
[专题系列] 美国总统经济顾问教你分析宏观数据 Extracting the Signal from the Noise


[论坛活动系列]
第一季翻译悬赏活动:《时代》2015年度风云人物!
第二季翻译悬赏活动: 见证呼吸化空气!
第三季翻译悬赏活动:《时代》2016年度风云人物!

使用道具

板凳
nicacc 在职认证  发表于 2016-12-25 09:25:37 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

thank you

使用道具

报纸
huhuhuhu 发表于 2016-12-25 09:46:36 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

谢谢分享

使用道具

地板
caifacai 发表于 2016-12-25 09:57:40 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

感谢分享好资源!圣诞快乐!

使用道具

7
ambut 发表于 2016-12-25 10:26:26 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

美知名对冲基金介绍深度学习(Deep Learning)最新研究成果

使用道具

8
franky_sas 发表于 2016-12-25 10:26:37 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

使用道具

9
vtmc 发表于 2016-12-25 10:30:26 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

2sigma NB得不行啊。感谢楼主的分享~~

使用道具

10
lm972 发表于 2016-12-25 10:30:27 |只看作者 |坛友微信交流群

回帖奖励 +1 个论坛币

谢谢分享

使用道具

您需要登录后才可以回帖 登录 | 我要注册

本版微信群
加好友,备注jr
拉您进交流群

京ICP备16021002-2号 京B2-20170662号 京公网安备 11010802022788号 论坛法律顾问:王进律师 知识产权保护声明   免责及隐私声明

GMT+8, 2024-4-20 06:15