Axial Attention in Multidimensional Transformers参考文献翻译
Ba, J. L., Kiros, J. R., & Hinton, G. E. (2016). Layer normalization. arXiv preprint arXiv:1607.06450.
中文翻译:Ba, J. L., Kiros, J. R., & Hinton, G. E. (2016). 层归一化. arXiv 预印本 arXiv:1607.06450.
Chen, X., Mishra, N., Rohaninejad, M., & Abbeel, P. (2018). PixelSNAIL: An improved autoregressive generative model. In International Conference on Machine Learning, pp. 863–871.
中文翻译:Chen, X., Mishra, N., Rohaninejad, M., & Abbeel, P. (2018). PixelSNAIL:一种改进的自回归生成模型. 在国际机器学习会议上,第 863–871 页.
Child, R., Gray, S., Radford, A., & Sutskever, I. (2019). Generating long sequences with sparse transformers. arXiv preprint arXiv:1904.10509.
中文翻译:Child, R., Gray, S., Radford, A., & Sutskever, I. (2019). 使用稀疏变换器生成长序列. arXiv 预印本 arXiv:1904.10509.
Ebert, F., Finn, C., Lee, A. X., & Levine, S. (2017). Self-supervised visual planning with temporal skip connections. In Conference on Robot Learning, pp. 344–356.
中文翻译:Ebert, F., Finn, C., Lee, A. X., & Levine, S. (2017). 使用时间跳跃连接的自监督视觉规划. 在机器人学习会议上,第 344–356 页.
Huang, Z., Wang, X., Huang, L., Huang, C., Wei, Y., & Liu, W. (2019). Ccnet: Criss-cross attention for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, pp. 603–612.
中文翻译:黄, Z., 王, X., 黄, L., 黄, C., 魏, Y., & 刘, W. (2019). Ccnet:用于语义分割的交叉注意力. 在 IEEE 国际计算机视觉会议论文集上,第 603–612 页.
Kalchbrenner, N., van den Oord, A., Simonyan, K., Danihelka, I., Vinyals, O., Graves, A., & Kavukcuoglu, K. (2017). Video pixel networks. In International Conference on Machine Learning, pp. 1771–1779.
中文翻译:Kalchbrenner, N., van den Oord, A., Simonyan, K., Danihelka, I., Vinyals, O., Graves, A., & Kavukcuoglu, K. (2017). 视频像素网络. 在国际机器学习会议上,第 1771–1779 页.
Kumar, M., Babaeizadeh, M., Erhan, D., Finn, C., Levine, S., Dinh, L., & Kingma, D. (2019). VideoFlow: A flow-based generative model for video. arXiv preprint arXiv:1903.01434.
中文翻译:Kumar, M., Babaeizadeh, M., Erhan, D., Finn, C., Levine, S., Dinh, L., & Kingma, D. (2019). VideoFlow:一种基于流的视频生成模型. arXiv 预印本 arXiv:1903.01434.
Menick, J., & Kalchbrenner, N. (2018). Generating high fidelity images with subscale pixel networks and multidimensional upscaling. arXiv preprint arXiv:1812.01608.
中文翻译:Menick, J., & Kalchbrenner, N. (2018). 使用子像素网络和多维上采样生成高保真图像. arXiv 预印本 arXiv:1812.01608.
Parikh, A. P., Täckström, O., Das, D., & Uszkoreit, J. (2016). A decomposable attention model for natural language inference. arXiv preprint arXiv:1606.01933.
中文翻译:Parikh, A. P., Täckström, O., Das, D., & Uszkoreit, J. (2016). 一种可分解的自然语言推理注意力模型. arXiv 预印本 arXiv:1606.01933.
Parmar, N., Vaswani, A., Uszkoreit, J., Kaiser, L., Shazeer, N., Ku, A., & Tran, D. (2018). Image transformer. In International Conference on Machine Learning, pp. 4052–4061.
中文翻译:Parmar, N., Vaswani, A., Uszkoreit, J., Kaiser, L., Shazeer, N., Ku, A., & Tran, D. (2018). 图像变换器. 在国际机器学习会议上,第 4052–4061 页.
Reed, S., van den Oord, A., Kalchbrenner, N., Colmenarejo, S. G., Wang, Z., Chen, Y., Belov, D., & de Freitas, N. (2017). Parallel multiscale autoregressive density estimation. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 2912–2921. JMLR. org.
中文翻译:Reed, S., van den Oord, A., Kalchbrenner, N., Colmenarejo, S. G., Wang, Z., Chen, Y., Belov, D., & de Freitas, N. (2017). 并行多尺度自回归密度估计. 在第 34 届国际机器学习会议论文集上,第 2912–2921 页.
Salimans, T., Karpathy, A., Chen, X., & Kingma, D. P. (2017). PixelCNN++: Improving the PixelCNN with discretized logistic mixture likelihood and other modifications. In International Conference on Learning Representations (ICLR).
中文翻译:Salimans, T., Karpathy, A., Chen, X., & Kingma, D. P. (2017). PixelCNN++:通过离散逻辑混合似然和其他改进提升 PixelCNN. 在国际学习表示会议上.
Theis, L., & Bethge, M. (2015). Generative image modeling using spatial LSTMs. In Advances in Neural Information Processing Systems, pp. 1927–1935.
中文翻译:Theis, L., & Bethge, M. (2015). 使用空间 LSTM 的生成图像建模. 在神经信息处理系统进展会议上,第 1927–1935 页.
Uria, B., Côté, M.-A., Gregor, K., Murray, I., & Larochelle, H. (2016). Neural autoregressive distribution estimation. The Journal of Machine Learning Research, 17(1), 7184–7220.
中文翻译:Uria, B., Côté, M.-A., Gregor, K., Murray, I., & Larochelle, H. (2016). 神经自回归分布估计. 机器学习研究杂志, 17(1), 7184–7220.
van den Oord, A., Kalchbrenner, N., & Kavukcuoglu, K. (2016a). Pixel recurrent neural networks. International Conference on Machine Learning (ICML).
中文翻译:van den Oord, A., Kalchbrenner, N., & Kavukcuoglu, K. (2016a). 像素递归神经网络. 国际机器学习会议.
van den Oord, A., Kalchbrenner, N., Vinyals, O., Espeholt, L., Graves, A., & Kavukcuoglu, K. (2016b). Conditional image generation with PixelCNN decoders. arXiv preprint arXiv:1606.05328.
中文翻译:van den Oord, A., Kalchbrenner, N., Vinyals, O., Espeholt, L., Graves, A., & Kavukcuoglu, K. (2016b). 使用 PixelCNN 解码器的条件图像生成. arXiv 预印本 arXiv:1606.05328.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). Attention is all you need. In Advances in Neural Information Processing Systems, pp. 5998–6008.
中文翻译:Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). 注意力就是一切. 在神经信息处理系统进展会议上,第 5998–6008 页.
Weissenborn, D., Täckström, O., & Uszkoreit, J. (2019). Scaling autoregressive video models. arXiv preprint arXiv:1906.02634.
中文翻译:Weissenborn, D., Täckström, O., & Uszkoreit, J. (2019). 扩展自回归视频模型. arXiv 预印本 arXiv:1906.02634.