forked from KEMT/zpwiki
		
	Update 'pages/students/2016/patrik_pavlisin/dp22/README.md'
This commit is contained in:
		
							parent
							
								
									81d53ed746
								
							
						
					
					
						commit
						8c85bda12a
					
				@ -108,10 +108,17 @@ Navrhovaný transformátor R sa skladá zo stohu rovnakých vrstiev. Každá vrs
 | 
			
		||||
## Zoznam použitej literatúry
 | 
			
		||||
 | 
			
		||||
[1]. VASWANI A., SHAZEER N., PARMAR N., USZKOREIT J., JONES L., GOMEZ N.A., KASIER L., POLUSUKHIN.I.: _Attention Is All You Need._ [online]. [citované 2017].
 | 
			
		||||
 | 
			
		||||
[2]. WANG Z., MA Y., LIU Z., TANG J.: _R-Transformer: Recurrent Neural Network Enhanced Transformer._ [online]. [citované 12-07-2019].
 | 
			
		||||
 | 
			
		||||
[3]. SRIVASTAVA S.: _Machine Translation (Encoder-Decoder Model)!._ [online]. [citované 31-10-2019].
 | 
			
		||||
 | 
			
		||||
[4]. ALAMMAR J.: _The Illustrated Transformer._ [online]. [citované 27-06-2018].
 | 
			
		||||
 | 
			
		||||
[5]. _Sequence Modeling with Neural Networks (Part 2): Attention Models_ [online]. [citované 18-04-2016].
 | 
			
		||||
 | 
			
		||||
[6]. GIACAGLIA G.: _How Transformers Work._ [online]. [citované 11-03-2019].
 | 
			
		||||
 | 
			
		||||
[7].  _Understanding LSMT Networks_ [online]. [citované 27-08-2015].
 | 
			
		||||
 | 
			
		||||
[8]. _6 Types of Artifical Neural Networks Currently Being Used in Machine Translation_ [online]. [citované 15-01-201].
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user